N. Marusov, I. Semenov

Similar documents
Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

The LHC Computing Grid

Distributing storage of LHC data - in the nordic countries

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

Andrea Sciabà CERN, Switzerland

Grid Computing. MCSN - N. Tonellotto - Distributed Enabling Platforms

The National Fusion Collaboratory

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

CC-IN2P3: A High Performance Data Center for Research

Grid Computing a new tool for science

AGIS: The ATLAS Grid Information System

The LHC Computing Grid

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

CernVM-FS beyond LHC computing

Storage Virtualization. Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan

Giovanni Lamanna LAPP - Laboratoire d'annecy-le-vieux de Physique des Particules, Université de Savoie, CNRS/IN2P3, Annecy-le-Vieux, France

ACCI Recommendations on Long Term Cyberinfrastructure Issues: Building Future Development

Grid Architectural Models

<Insert Picture Here> Enterprise Data Management using Grid Technology

Outline. Infrastructure and operations architecture. Operations. Services Monitoring and management tools

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

Grid Scheduling Architectures with Globus

Summary of the LHC Computing Review

Lessons Learned in the NorduGrid Federation

The Grid: Processing the Data from the World s Largest Scientific Machine

ONE-ITS Science Gateway

A Simulation Model for Large Scale Distributed Systems

Network Needs of US-China Fusion Research Collaborations

Scientific data management

SUPPORTING EFFICIENT EXECUTION OF MANY-TASK APPLICATIONS WITH EVEREST

The EGEE-III Project Towards Sustainable e-infrastructures

The CMS Computing Model

Travelling securely on the Grid to the origin of the Universe

Grid Computing at the IIHE

Grid Computing: dealing with GB/s dataflows

Grid Computing. Lectured by: Dr. Pham Tran Vu Faculty of Computer and Engineering HCMC University of Technology

IEPSAS-Kosice: experiences in running LCG site

DALA Project: Digital Archive System for Long Term Access

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

An Experimental Cloud Resource Broker System for Virtual Application Control with VM Allocation Scheme

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

Monte Carlo Production on the Grid by the H1 Collaboration

2014 年 3 月 13 日星期四. From Big Data to Big Value Infrastructure Needs and Huawei Best Practice

Introduction to Grid Technology

Interoperating AliEn and ARC for a distributed Tier1 in the Nordic countries.

Overview of ATLAS PanDA Workload Management

The Virtual Observatory and the IVOA

Philippe Laurens, Michigan State University, for USATLAS. Atlas Great Lakes Tier 2 collocated at MSU and the University of Michigan

Distributed e-infrastructures for data intensive science

Next-generation IT Platforms Delivering New Value through Accumulation and Utilization of Big Data

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

RADU POPESCU IMPROVING THE WRITE SCALABILITY OF THE CERNVM FILE SYSTEM WITH ERLANG/OTP

Creating synergy through private cloud

Grid Computing: dealing with GB/s dataflows

Computing at the Large Hadron Collider. Frank Würthwein. Professor of Physics University of California San Diego November 15th, 2013

Open data and scientific reproducibility

irods for Data Management and Archiving UGM 2018 Masilamani Subramanyam

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

The JINR Tier1 Site Simulation for Research and Development Purposes

Distributed Computing Framework. A. Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille

HEP Grid Activities in China

Geographically Distributed Software Defined Storage (the proposal)

Chapter 2 Introduction to the WS-PGRADE/gUSE Science Gateway Framework

MONTE CARLO SIMULATION FOR RADIOTHERAPY IN A DISTRIBUTED COMPUTING ENVIRONMENT

EarthCube and Cyberinfrastructure for the Earth Sciences: Lessons and Perspective from OpenTopography

Network Based Hard/Soft Information Fusion Network Architecture/SOA J. Rimland

Introduction to Grid Computing

Grid Computing Activities at KIT

Status of KISTI Tier2 Center for ALICE

e-infrastructures in FP7 INFO DAY - Paris

Hands-on tutorial on usage the Kepler Scientific Workflow System

. Smart-Cities and Cloud Computing. Panel Discussion

Postgres Plus and JBoss

Virtualizing a Batch. University Grid Center

First Experience with LCG. Board of Sponsors 3 rd April 2009

Review of the Compact Muon Solenoid (CMS) Collaboration Heavy Ion Computing Proposal

Modeling User Submission Strategies on Production Grids

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

WLCG Transfers Dashboard: a Unified Monitoring Tool for Heterogeneous Data Transfers.

Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21)

High Throughput WAN Data Transfer with Hadoop-based Storage

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

HEP replica management

Distributed Data Management with Storage Resource Broker in the UK

Computing grids, a tool for international collaboration and against digital divide Guy Wormser Director of CNRS Institut des Grilles (CNRS, France)

AMAZON S3 FOR SCIENCE GRIDS: A VIABLE SOLUTION?

European Grid Infrastructure

Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland

A REMOTE CONTROL ROOM AT DIII-D

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

GC-APWG Global Changing Asian-Pacific Wide Grid. Dr. Guoqing Li, CEODE/CAS 25,May,2008 UN GAID E-SDDC, Shanghai

Grid and Cloud Activities in KISTI

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

Data Intensive processing with irods and the middleware CiGri for the Whisper project Xavier Briand

GA A26400 CUSTOMIZABLE SCIENTIFIC WEB-PORTAL FOR DIII-D NUCLEAR FUSION EXPERIMENT

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

EUDAT & SeaDataCloud

Metadata Models for Experimental Science Data Management

EUDAT-B2FIND A FAIR and Interdisciplinary Discovery Portal for Research Data

South African Science Gateways

Transcription:

GRID TECHNOLOGY FOR CONTROLLED FUSION: CONCEPTION OF THE UNIFIED CYBERSPACE AND ITER DATA MANAGEMENT N. Marusov, I. Semenov Project Center ITER (ITER Russian Domestic Agency N.Marusov@ITERRF.RU)

Challenges of ITER data usage Data accumulation & distribution ~5 PB/year, ~100 PB for a lifetime Huge amount of data cannot be analyzed by humans Automated conversion of raw data to knowledge Scientific collaboration all over the world Diversity of end-user needs and tools Future integration of ITER scientific achievements with other TOKAMAKs 2

ITER research as control process Fusion community formulates the goals and gets the results Fusion Community Knowledge Information circulating in control loops represents experimental data and knowledge Goals Knowledge Management System Fusion Technology Results Impact to technology is done trough the Knowledge Management System Data ITER machine Fusion technology is viewed as the controlled object ITER machine work as sensors to assess the quality of technology 3

Domain knowledge allocation Slow data processing (10 min - hours) Fast data processing (< 10 min) 4

Unified cyberspace (UC) concept Knowledge Management System (KMS) is one of key components of ITER computing model Distributed nature of ITER collaboration Community-oriented research workflow Liaison standards for interoperation in heterogeneous computing environments Machine learning-enabled environment Distributed community-driven KMS semantically linking diverse computing resources, modeling codes and data storages on the base of existing standards 5

Most of such problems are already solved within existing projects Worldwide LHC Computing Grid (CERN + 42 countries) National Fusion Collaboratory Project (USA) European Grid Infrastructure European Middleware Initiative based on well-known concepts such as e-science, Science 2.0, Cyberinfrastructure and Semantic Web 6

General view of UC infrastructure Wiki-sites & Social Networks Virtual Laboratory User s Applications High-level Services SaaS Application Data Providers Computing Resources Experimen tal Facilities Applications Information Services Network VO Web Portal Storage Personal Environment Workflows Basic Grid Services Storage Element Computing Element End User Robotic Agents Other 7

Basic Grid middleware Security / Virtual organizations management service User Interface Computing element Storage element Information services Workload management 8

Data accumulation & distribution - LHC Worldwide LHC Computing Grid Tier 0 (CERN) Data recording Initial data reconstruction Data distribution Tier 1 (11 + KISTI, Korea in progress) Permanent storage Re-processing Analysis 10 Gbit/s links Tier 2 (~150 centres) Simulation End-user analysis Tier 3 (End-users) Customized processing Overall ~160 sites, 39 countries 300,000 cores 200 PB of storage 2 million jobs/day 9

Data accumulation & distribution - ITER ITER Site Tier 0 Full replication Domestic Agency Domestic Agency Domestic Agency Tier 1 National Institute National Institute Processed Sample or partial replica Tier 2 Processed Sample End User Tier 3 10

Unified Data providers Metadata update Information Service Facility-independent Metadata scheme Cache Storage Native data format On demand data selection Sampling processor Server-side data manipulation HTTP MDSPlus End User Unified Data Provider 11

Fusion research poses a lot of loosely-tied tasks Parametric optimization of fusion facilities and algorithms Data mining & scaling laws Empirical dynamic models generation 3D-animation of dynamic processes Monte-Carlo & Ray tracing 12

Job flow in the WLCG/EGEE Grid 13

Grid-enabled applications as Services Direct Grid API usage Direct Service API usage SaaS Application 1 Grid interface End User Workflow composer SaaS Application 2 Grid interface Grid Workflow sharing SaaS Application 3 Grid interface 14

Cooperation within ITER & ITER as a partner of cooperation JET ITER KSTAR T-15M Fusion community cyberspace DIII-D DEMO Step 1: Prove feasibility of Worldwide Fusion Laboratory on the base of ITER and existing solutions. Step 2: Integrate storages of other TOKAMAKs with the unified cyberspace on the base of proven technology. Step 3: New super-collaboration endeavors for DEMO optimized fusion power plant. 15

Thank you for your attention! Discussion 16