GRID 2008 International Conference on Distributed computing and Grid technologies in science and education 30 June 04 July, 2008, Dubna, Russia

Similar documents
GRIDS INTRODUCTION TO GRID INFRASTRUCTURES. Fabrizio Gagliardi

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Grid Computing a new tool for science

High Performance Computing from an EU perspective

PROJECT FINAL REPORT. Tel: Fax:

Tier-2 structure in Poland. R. Gokieli Institute for Nuclear Studies, Warsaw M. Witek Institute of Nuclear Physics, Cracow

CLOUDS OF JINR, UNIVERSITY OF SOFIA AND INRNE JOIN TOGETHER

The LHC Computing Grid

Travelling securely on the Grid to the origin of the Universe

Introduction to FREE National Resources for Scientific Computing. Dana Brunson. Jeff Pummill

The EGEE-III Project Towards Sustainable e-infrastructures

Andrea Sciabà CERN, Switzerland

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

Introduction to Grid Infrastructures

IEPSAS-Kosice: experiences in running LCG site

e-infrastructures in FP7 INFO DAY - Paris

The LHC Computing Grid

e-infrastructure: objectives and strategy in FP7

The Grid: Processing the Data from the World s Largest Scientific Machine

The EuroHPC strategic initiative

Europe and its Open Science Cloud: the Italian perspective. Luciano Gaido Plan-E meeting, Poznan, April

The PICTURE project, ICT R&I priorities in EaP, areas of cooperation

DG CONNECT (Unit H5) Update on Data Centre Activities

The EPIKH, GILDA and GISELA Projects

Federated E-infrastructure Dedicated to European Researchers Innovating in Computing network Architectures

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

einfrastructures Concertation Event

The Research Infrastructures Programe and its NCP Network RICH

Outline. Infrastructure and operations architecture. Operations. Services Monitoring and management tools

Storage and I/O requirements of the LHC experiments

NEWCOM# The Network of Excellence in Wireless COMmunications

High-Performance Computing Europe s place in a Global Race

Digitising European industry

Moving e-infrastructure into a new era the FP7 challenge

EuroHPC Bologna 23 Marzo Gabriella Scipione

EuroHPC and the European HPC Strategy HPC User Forum September 4-6, 2018 Dearborn, Michigan, USA

Advanced School in High Performance and GRID Computing November Introduction to Grid computing.

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

Integration of Network Services Interface version 2 with the JUNOS Space SDK

AMRES Combining national, regional and & EU efforts

UNICORE Globus: Interoperability of Grid Infrastructures

Information and Communication Technologies (ICT) thematic area

Grid Computing at Ljubljana and Nova Gorica

Batch Services at CERN: Status and Future Evolution

The Virtual Observatory and the IVOA

Supercomputer and grid infrastructure! in Poland!

Grid and Cloud Activities in KISTI

Pre-announcement of upcoming procurement, NWP18, at National Supercomputing Centre at Linköping University

A European Vision and Plan for a Common Grid Infrastructure

ISTITUTO NAZIONALE DI FISICA NUCLEARE

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

The European DataGRID Production Testbed

Garuda : The National Grid Computing Initiative Of India. Natraj A.C, CDAC Knowledge Park, Bangalore.

Presentation Title. Grid Computing Project Officer / Research Assistance. InfoComm Development Center (idec) & Department of Communication

Opportunities A Realistic Study of Costs Associated

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

EGEODE. !Dominique Thomas;! Compagnie Générale de Géophysique (CGG, France) R&D. Expanding Geosciences On Demand 1. «Expanding Geosciences On Demand»

EuroHPC: the European HPC Strategy

TEDS BEE TEST OF DIGITAL SERVICES FOR BUILDING ENERGY EFFICIENCY A PROJECT TOWARDS THE ENERGY SAVING AND SUSTAINABILITY OF PUBLIC BUILDINGS

Huawei European Research University Partnerships. Michael Hill-King European Research Institute, Huawei

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

Some aspect of research and development in ICT in Bulgaria. Authors Kiril Boyanov and Stefan Dodunekov

Smart Grid Objective 6.1 in WP2013

First Experience with LCG. Board of Sponsors 3 rd April 2009

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

EPOS a long term integration plan of research infrastructures for solid Earth Science in Europe

ESFRI Strategic Roadmap & RI Long-term sustainability an EC overview

CESGA: General description of the institution. FINIS TERRAE: New HPC Supercomputer for RECETGA: Galicia s Science & Technology Network

Striving for efficiency

Experience of the Tunisian Participation to the EU s Horizon 2020 Framework Programme as an Associated Country

SZDG, ecom4com technology, EDGeS-EDGI in large P. Kacsuk MTA SZTAKI

H2020 EUB EU-Brazil Research and Development Cooperation in Advanced Cyber Infrastructure. NCP Training Brussels, 18 September 2014

Research Infrastructures and Horizon 2020

EGI-Engage: Impact & Results. Advanced Computing for Research

Computing for LHC in Germany

CYFRONET SITE REPORT IMPROVING SLURM USABILITY AND MONITORING. M. Pawlik, J. Budzowski, L. Flis, P. Lasoń, M. Magryś

ATLAS NorduGrid related activities

HPC IN EUROPE. Organisation of public HPC resources

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

Petaflop Computing in the European HPC Ecosystem

Project overview. Gelsomina Pappalardo National Research Council of Italy. AEROCOM - ACTRIS Meeting Hamburg, Germany 26 September 2013

Scope of activities scientific and research work in informatics, information technology, control theory, robotics and artificial intelligence

Media (NEM) Initiative

FEDERICA Federated E-infrastructure Dedicated to European Researchers Innovating in Computing network Architectures

RESEARCH INFRASTRUCTURES IN HORIZON Marine Melkonyan National University of S&T MISIS

South African Science Gateways

Italy - Information Day: 2012 FP7 Space WP and 5th Call. Peter Breger Space Research and Development Unit

Computing grids, a tool for international collaboration and against digital divide Guy Wormser Director of CNRS Institut des Grilles (CNRS, France)

The European Platform in Network and Information Security (NIS) Fabio Martinelli

EGI: Linking digital resources across Eastern Europe for European science and innovation

EUDAT & SeaDataCloud

Challenges in the field of Research Infrastructures, ESFRI

CC-IN2P3: A High Performance Data Center for Research

The Future of Solid State Lighting in Europe

The grid for LHC Data Analysis

HPC & Quantum Technologies in Europe

Assessment of the progress made in the implementation of and follow-up to the outcomes of the World Summit on the Information Society

Grids and Security. Ian Neilson Grid Deployment Group CERN. TF-CSIRT London 27 Jan

Grid Challenges and Experience

Easy Access to Grid Infrastructures

Transcription:

GRID 2008 International Conference on Distributed computing and Grid technologies in science and education 30 June 04 July, 2008, Dubna, Russia VO of EIMM within EGEE GRID: sharing of computing resources Grzegorz Musiał Faculty of Physics, Adam Mickiewicz University Ul. Umultowska 85, 61-614 Poznań, POLAND E-mail: gmusial@amu.edu.pl http://www.staff.amu.edu.pl/~g.musial/ G. Musiał, GRID 2008 1

Coauthors: Grzegorz Kamieniarz, Bogusz Błaszkiewicz Faculty of Physics, Adam Mickiewicz University, Poland All authors are also members of the European Institute of Molecular Magnetism (EIMM) G. Musiał, GRID 2008 2

Outline 1. Motivation 2. NoE MAGMANet EIMM + ISG 3. Integrated Laboratory Modeling and Computing 4. VO ILMC sharing of computing resources 5. Conclusions G. Musiał, GRID 2008 3

Motivation Single-molecule magnets the basis of the next generation of data storage media or quantum computers. The aim of this talk is to present sharing of computing resources in the recently founded European Institute of Molecular Magnetism (EIMM). The computational problems need effective resources shared among the partners involved. According to this paradigm, a grid-based large-scale computing environment offers such an efficient mutual use of computing resources. G. Musiał, GRID 2008 4

Motivation Computing of energy spectrum would take more than half of a year on a single CPU. The dodecanuclear Ni 12 wheel [Ni 12 (O 2 CMe) 12 (chp) 12 (H 2 O) 6 (THF) 6 ] Ni blue S=1 The higher S value or the larger molecule the larger computing problem with diagonalization of a Hamiltonian (S-mixing, anisotropy included!). G. Musiał, GRID 2008 5

Project MAGMANet EIMM The European Commission 10.7M of funding from 6 th FP Network of centres of Excellence, NoE MAGMANet started May 1 st, 2005 for 4 years Project MAGMANet, Molecular Approach to Nanomagnets and Multifunctional Materials, coordinated by the Italian INSTM (Consorzio Interuniversitario Nazionale per la Scienza e Tecnologia dei Materiali), Florence, prof. Dante Gatteschi started April 12 th, 2008 European Institute of Molecular Magnetism located in INSTM in Florence G. Musiał, GRID 2008 6

PL-Node, AMU Poznań European Institute of Molecular Magnetism Researchers from many cities are focused in 15 nodes G. Musiał, GRID 2008 7

Project MAGMANet EIMM Important goals of molecular magnetism in the last two decades to provide examples of novel phenomena like: magnetic hysteresis at the molecular level, quantum effects in the dynamics of the macroscopic magnetization, photoswitchable magnetic materials, polyfunctional materials combining conduction and bulk ferromagnetism. Among others single-molecule magnets should become the basis of the next generation of data storage media or transistors in new generation processors. G. Musiał, GRID 2008 8

EIMM The EIMM will continue to integrate European research in Molecular Magnetism by bringing together the scientific excellence of the partners and providing access to world class instrumentation. We hope we are above the critical mass to achieve our goals. The institute will provide the first European Doctorate in molecular magnetism allowing students access to a range of research facilities within the partner institutions. G. Musiał, GRID 2008 9

EIMM Industrial Support Group Industrial players with an interest in molecular magnetism and a need to maintain their competitive edge may join the EIMM s Industrial Support Group (ISG) with Pietro Perlo (Centro Ricerche Fiat) the Chairman and Fernando Palacio (Zaragoza University) the executive Secretary. Pietro The ISG organizes a range of informative and technology transfer activities providing opportunities for collaborative projects. It is free of charge and open to all: http://magmanet.unizar.es/magmanet-isg/ Fernando G. Musiał, GRID 2008 10

EIMM ISG A synergic effort coming from synthetic chemists, experimental physicists and theoreticians to move towards applications and integration with the industrial world. Everybody is cordially invited. Apart from world class instrumentation platforms the network defines specialized (virtual) integrated laboratories to achieve integration, research and spreading of excellence objectives. G. Musiał, GRID 2008 11

Integrated Laboratory Modeling and Computing Two possible meanings of integrated laboratory: a virtual existence, with no material existence, a set of experimental/theoretical specialized labs sharing expertise on a complementary basis and offering their experience to solve problems for other participants. 10 integrated laboratories is set up within MAGMANet, among others Modeling and Computing, headed by prof. G. Kamieniarz. This laboratory is of the 1 st and of the 2 nd type, as sharing the computing basis is natural within grids. G. Musiał, GRID 2008 12

Sharing of computing resources EGEE project aims at providing researchers in academia and industry with access to major computing resources independent of their geographic location. To organize sharing of computing resources, we have contacted our partner at the Poznań Supercomputing and Networking Centre involved in the EGEE Grid project, also funded by the EC. Support for VOs: ATLAS, ALICE, LHCB, CMS, VOCE, balticgrid, gaussian. G. Musiał, GRID 2008 13

Sharing of computing resources We have integrated the resources of five partner nodes in EIMM: INSTM, CNR-INFM from Italy, AMU from Poland, UVEG from Spain, UAIC from Romania (together about 5 TFlops) to start the VO ILMC within the EGEE grid. Our largest computer cluster at the UAIC node, Romania, 130 quad-core CPUs. We also obtain cross-linking of important EU projects. G. Musiał, GRID 2008 14

Sharing of computing resources The broad knowledge base provided by this Integrated Laboratory affords transfer of organization skills, access to variety of specialized computational tools and software as well as comprehensive expertise which can address problems in the following important areas: high performance scientific computing in materials science, grid computing for public and commercial units, e-learning management, optimization tasks based on genetic algorithms, artificial intelligence in bio-medicine. G. Musiał, GRID 2008 15

Sharing of computing resources To make the decisions easier we argue: sharing means not loosing control of own resources, as one can offer only some percent of computing power for a grid, instead of it, one can run much bigger jobs. For example: one of our Italian partners concluded that it would be worth to join our VO if they obtain access to the cluster with nodes having 64 GB RAM. What is the answer? Our VO obtained the access to some resources within Baltic Grid where there are nodes of 64 GB RAM! G. Musiał, GRID 2008 16

Sharing of computing resources Offering only some percent of computing power for a grid, one obtains the access to the specialized computational resources and software. Nodes Local switch Firewall/ router Switch Backbone network Access node Internet Monitoring for whole system Access node Nodes Any cluster Any our partners, scientific or industrial, connection through the internet, Virtual Organization in a grid, Globus Toolkit, authentication + authorization G. Musiał, GRID 2008 17

PL-Grid Polish Grid Initiative The Polish Grid Consortium (2007) a response to the needs of Polish scientists, compatible with ongoing grid activities in other European countries (within EGI) and all over the world, made up of 5 largest Polish supercomputing & networking centers (to be financed by national funds for 2008-2010). The main aim of the PL-Grid Project is to create and to develop a stable Polish Grid infrastructure, fully compatible and interoperable with European and worldwide Grids. The Project should provide scientific communities in Poland with Grid services, enabling realization of the e-science model in various scientific fields. Thanks to prof. H.F. Hoffmann G. Musiał, GRID 2008 18

Conclusions Organization and software problems for sharing of computing resources are solved already, there are only the problems to transfer the knowledge how to get and to run parallel applications, the political will and of a mentality of (wo)men. Most of calculations within molecular magnetism scale very well with efficiency about 0.9. Many of the simulations can be run with hundreds or even thousands of parallel processes. Not only parallel processing but grid computing opens a new scale of modeling in molecular magnets. G. Musiał, GRID 2008 19

Conclusions Scalability of simulations is better and better at larger and larger molecules. This makes possible to consider much larger systems. The larger the systems considered the better the analysis the more reliable the results. The candidates for the next generation of data storage media or transistors in new generation processors are rather large molecules for which we utilize self-assembling and self-organizing when putting them on the surface. Thus, modeling them needs a huge number of CPUs available at grid computing. G. Musiał, GRID 2008 20

Acknowledgements The work was supported in part by the European Commission under the project MAGMANet NMP3-CT-2005-515767 Some cluster and grid organization details come from the Poznań Supercomputing and Networking Center Thanks for your attention and patience!!! E-mail: gmusial@amu.edu.pl http://www.man.poznan.pl/ EIMM: http://www.eimm.org/ ISG: http://magmanet.unizar.es/magmanet-isg/ G. Musiał, GRID 2008 21