SAM at CCIN2P3 configuration issues

Similar documents
D0 Grid: CCIN2P3 at Lyon

Benoit DELAUNAY Benoit DELAUNAY 1

LCG data management at IN2P3 CC FTS SRM dcache HPSS

Wide-Area Networking at SLAC. Warren Matthews and Les Cottrell (SCS Network Group) Presented at SLAC, April

Grid Computing Activities at KIT

KLOE software on Linux

Distributed Monte Carlo Production for

(Tier1) A. Sidoti INFN Pisa. Outline: Tasks and Goals The analysis (physics) Resources Needed

CC-IN2P3 activity. irods in production: irods developpements in Lyon: SRB to irods migration. Hardware setup. Usage. Prospects.

The EU DataGrid Testbed

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

Table 9. ASCI Data Storage Requirements

Tier2 Centre in Prague

The ATLAS Tier-3 in Geneva and the Trigger Development Facility

Austrian Federated WLCG Tier-2

High-Energy Physics Data-Storage Challenges

Globus Online and HPSS. KEK, Tsukuba Japan October 16 20, 2017 Guangwei Che

Data Transfers Between LHC Grid Sites Dorian Kcira

CASTORFS - A filesystem to access CASTOR

Data transfer over the wide area network with a large round trip time

Parallel Storage Systems for Large-Scale Machines

Fermilab FERMILAB-Conf-03/207-E August 2003

1.Remote Production Facilities

10 Gbit/s Challenge inside the Openlab framework

BaBar Computing: Technologies and Costs

The CMS Computing Model

NAREGI Middleware Mediator

High-density Grid storage system optimization at ASGC. Shu-Ting Liao ASGC Operation team ISGC 2011

SPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2

Andy Kowalski Ian Bird, Bryan Hess

NCAR Globally Accessible Data Environment (GLADE) Updated: 15 Feb 2017

designed. engineered. results. Parallel DMF

Storage Resource Sharing with CASTOR.

Experimental Computing. Frank Porter System Manager: Juan Barayoga

irods usage at CC-IN2P3 Jean-Yves Nief

A data Grid testbed environment in Gigabit WAN with HPSS

Database Services at CERN with Oracle 10g RAC and ASM on Commodity HW

The INFN Tier1. 1. INFN-CNAF, Italy

Linux Compute Cluster in the German Automotive Industry

High Performance Computing Course Notes High Performance Storage

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss

CMS Belgian T2. G. Bruno UCL, Louvain, Belgium on behalf of the CMS Belgian T2 community. GridKa T1/2 meeting, Karlsruhe Germany February

Edinburgh (ECDF) Update

Long Term Data Preservation for CDF at INFN-CNAF

Parallel File Systems. John White Lawrence Berkeley National Lab

DØ Southern Analysis Region Workshop Goals and Organization

Distributed ASCI Supercomputer DAS-1 DAS-2 DAS-3 DAS-4 DAS-5

SPECIFICATION FOR NETWORK ATTACHED STORAGE (NAS) TO BE FILLED BY BIDDER. NAS Controller Should be rack mounted with a form factor of not more than 2U

Earth Observation Innovation Platform _. Price List

Data oriented job submission scheme for the PHENIX user analysis in CCJ

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

Distributing storage of LHC data - in the nordic countries

Towards Network Awareness in LHC Computing

400GB GB

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Scaling PortfolioCenter on a Network using 64-Bit Computing

High bandwidth, Long distance. Where is my throughput? Robin Tasker CCLRC, Daresbury Laboratory, UK

CERN and Scientific Computing

The JANUS Computing Environment

Implementing a Digital Video Archive Based on the Sony PetaSite and XenData Software

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

A Grid Computing Centre at Forschungszentrum Karlsruhe. Response on the Requirements for a Regional Data and Computing Centre in Germany 1 (RDCCG)

Milestone Systems Husky Network Video Recorders Product Comparison Chart September 9, 2016

Tier-2 structure in Poland. R. Gokieli Institute for Nuclear Studies, Warsaw M. Witek Institute of Nuclear Physics, Cracow

IBM řešení pro větší efektivitu ve správě dat - Store more with less

Data Block. Data Block. Copy A B C D P HDD 0 HDD 1 HDD 2 HDD 3 HDD 4 HDD 0 HDD 1

Figure 1: cstcdie Grid Site architecture

Influence of Distributing a Tier-2 Data Storage on Physics Analysis

Storage Systems. Storage Systems

Introduction to CARC. To provide high performance computing to academic researchers.

Storage and I/O requirements of the LHC experiments

Introduction to High Performance Parallel I/O

Data oriented job submission scheme for the PHENIX user analysis in CCJ

Resource Management on a Mixed Processor Linux Cluster. Haibo Wang. Mississippi Center for Supercomputing Research

ATLAS Tier-3 UniGe

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

Accelerating Parallel Analysis of Scientific Simulation Data via Zazen

Reprocessing DØ data with SAMGrid

Technical Brief: Specifying a PC for Mascot

A copy can be downloaded for personal non-commercial research or study, without prior permission or charge

SONAS Best Practices and options for CIFS Scalability

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

UW-ATLAS Experiences with Condor

Mass Storage at the PSC

The DETER Testbed: Overview 25 August 2004

Scalability and Classifications

Storage and Storage Access

Managing Petabytes of data with irods. Jean-Yves Nief CC-IN2P3 France

dcache tape pool performance Niklas Edmundsson HPC2N, Umeå University

HPC Workshop. Nov. 9, 2018 James Coyle, PhD Dir. Of High Perf. Computing

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster

The TinyHPC Cluster. Mukarram Ahmad. Abstract

Solutions for iseries

Massive Data Processing on the Acxiom Cluster Testbed

HITACHI VIRTUAL STORAGE PLATFORM G SERIES FAMILY MATRIX

EU DataGRID testbed management and support at CERN

irods usage at CC-IN2P3: a long history

Securing a HENP Computing Facility

Tier2 Centers. Rob Gardner. University of Chicago. LHC Software and Computing Review UC San Diego Feb 7-9, 2006

Transcription:

SAM at CCIN2P3 configuration issues Patrice Lebrun - IPNL/IN2P3

CCIN2P3 present actions Computing and data storage services for about 45 experiments Regional Center services for: EROS II BaBar ( Tier A) D0... AUGER LHC

CCIN2P3 services

Linux Workers 16 GB of SCSI disk/worker with 2 GB/CPU for scratch Fast Internet 100 Mb/s Linux (RedHat 6.1) -> 96 dual PIII 750MHz + 100 dual PIII 1GHz From Now Half of them are upgraded to RedHat 7.2 28 dual PIII 1.3 GHz will be installed on September 2002

Networking Academic public network Renater 2 based on virtual networking (ATM) with guaranteed bandwidth (VPN on ATM) Lyon CERN at 155 Mb Lyon US is going through CERN Bottleneck FNAL Chicago ( ~ 30 Mb/s ) Lyon STARtap at 100 Mb, STARtap to Esnet at 50 Mb (for BaBar).

HPSS (High Performance Storage System) HPSS local developments: Interface with RFIO: API: C, Fortran (via cfio) API: C++ (iostream) (for gcc and KCC) Development of RFIO 64bits will be implemented in HPSS bbftp secure parallel ftp using RFIO interface HPSS and D0 800 GB of Cache disk (could be larger in future) About 8 TB stored (MC and Imported Data)

SAM Stations Two SAM stations Dual PIII 1 GHz with 1 GB of memory RedHat 6.1 Administrators: L. Duflot and P. Lebrun Version: v4_1_0 Flavor: Linux+2.2 ccin2p3-analysis (general purpose) ccin2p3-production (dedicated to MC production) Used to store files at FNAL bbftp recompiled at Lyon with RFIO Actual limit of space for SAM cache is 10 GB and not seen by other computers. (Not used)

SAM - HPSS Issues IP 100 Mb/s FNAL bbftp GB Ethernet 2 Get file name location 2 GB Client Job Worker > 400 CPU 1 SAM (ccd0) Sam locate Rfstream (iostream) 6 Local mount 10 GB Copy file (RFIO) 3 rfcp Locate file on tape (30s) 4 HPSS cchpssd0:// Tape 5 TB Robot stk 3490 20 GB Copy file to HPSS s cache (12 MB/s)

Comments Advantages low cost (man power, hardware...) efficiency (high number of workers and jobs) low maintenance (no daemon running (SAM) and/or no mount (NFS) on each worker) Disadvantages No optimization performed by SAM but only by HPSS (no physics oriented) any other?...

Some words about SAM Store is done Copy file from HPSS on local disk byrfcp rfcp cchpssd0.in2p3.fr:/hpss/in2p3.fr/group/d0/mcprod/rec_filename /d0cache/import/ cp /<path>/import_description_file /d0cache/import cd /d0cache/import sam store --source=. --descrip=import_description_file rm -f * (clean buffer) If SAM gives only the file location BBFTP can read file directly from HPSS Problem with the Import description file which has to be on the same directory where the storing file is! To SAM experts : is true or false?

D0 code is not touched: Implementation in rfstream class only Mode 1 (copy on local disk) check space on local disk rfcp (rfio copy) from HPSS and open file locally close and remove file from disk Mode 2 Work to do open and read file using API directly from HPSS (but 5 times slower than mode 1) SAM requirement has to return only the file name location without any check on the file (performed by HPSS) when file is located in HPPS (cchpssd0 server) Exportation from FNAL to CCIN2P3 has to be investigated

DATA GRID Testbed is involved in the frame of WP6 (DataGRID) Participation in WPx groups (x=8,7,9,10) Integration of BQS batch system into Globus Batch Workers User Home Laboratory Globus Gatekeeper Batch Scheduler BQS. Computing Element HPSS Xtage Storage Element IN2P3 Computing Center

Conclusion D0 code not changed only ccin2p3 iostream are modified SAM has to see cchpssd0 (HPSS server) like a local disk. Files should be accessed locally after an RFIO copy (mode 1) or open and read using API (mode 2) Need help from a SAM expert Hope to do some tests in September SAM with Data Grid should be also tested (D0GRID) at the CCIN2P3