NCP Computing Infrastructure & T2-PK-NCP Site Update. Saqib Haleem National Centre for Physics (NCP), Pakistan
|
|
- Violet Murphy
- 5 years ago
- Views:
Transcription
1 NCP Computing Infrastructure & T2-PK-NCP Site Update Saqib Haleem National Centre for Physics (NCP), Pakistan
2 Outline NCP Overview Computing Infrastructure at NCP WLCG T2 Site status Network status and Issues Conclusion and Future work
3 About NCP The National Centre for Physics (NCP), Pakistan has been established to promote research in Physics & applied disciplines in the country. Collaboration with many in International organizations including CERN, SEASME, ICTP, TWAS Major Research Programs: Experimental High Energy Physics, Theoretical and Plasma Physics, Nano Sciences and Catalysis, Laser Physics, vacuum Sciences and technology, Earthquake studies. Organize/hosts International & National workshops, conferences every year. International Scientific School (ISS), NCP school on LHC Physics.
4 Computing Infrastructure Overview Installed Resources in Data Centre No. of Physical Servers 89 Processor sockets 140 CPU Cores 876 Memory Capacity 1.7TB No. of Disks ~600 Storage Capacity Network Equipment (Switches, Router, Firewalls) ~500TB ~50
5 Computing Infrastructure Overview Computating Infrastructure of NCP is mainly utilized for two projects: Major Portion of compute and storage resources are being used in WLCG T2 site Project & Local Batch system focusing Experimental High Energy Physicist community Rest Allocated to High Performance Computing (HPC) cluster.
6 WLCG Site at NCP status NCP-LCG2 site hosted at National Centre for Physics (NCP), for WLCG CMS experiment. Site was established in 2005, with very limited resources. Later on resources upgraded in The Only CMS Tier-2 Site in Pakistan( T2_PK_NCP) NCP-LCG2 WLCG resource Pledges Installed Pledged WLCG Resource Summary Physical Hosts 63 CPU sockets 126 No. of CPU Cores 524 HEPSPEC Disk Storage Capacity Network Connectivity Compute Elements (CE) DPM Storage Disk Nodes 330 TB 1 Gbps (shared) 03 02( CREAM-CE) 01 (HTCONDOR-CE) 13 CPU (HEP-SPEC06) Disk (TeraBytes) *
7 WLCG Hardware Status Computing Servers Hardware Sun Fire X4150(Intel(R) Xeon(R) CPU 3.16GHz) Dell Power Edge R610 (Intel(R) Xeon(R) CPU 2.93GHz) Sockets /cores Qty. 02/ / Total Cores 524 cores Storage Servers Storage Server Quantity Usable Capacity Transtec Lynx TB x1tb Dell Power Edge T X 2TB Existing Hardware is very old, and replacement of hardware is in progress. Recent Purchase: Compute Servers : 3 X Dell PowerEdge 740R (2 x Intel Xeon Silver 4116 Processor) 116 CPU cores 1 X Lenovo systemx 3650 (2 x Intel Xeon E52699v4 Processor) Storage: Dell PowerEdge 740xd CPU (2 x Intel Xeon Silver 4116 Processor) = 16 cores (50TB) Dell PowerEdge R430 CPU 2 x 2609v4 = 16 cores (100TB PV-MD1200) More Procurement is in process TB 150TB Raw
8 NCP-LCG2 Site Availability & Reliability
9 Usage Statistics
10 HPC Cluster HPC cluster Deployed is available based on Rocks Linux Distribution, and available 24x CPU cores are available. Applications Installed WEIN2K Suspect GEANT2 R/RMPI Quantum Espresso Scotch MCR.. Material Science Climate Modeling, Earthquake studies Research Areas Condensed Matter Physics Plasma Physics.. Cluster Usage ---- Last 2 years Cluster Usage Trend---- Last 3 years
11 Computing Model Partially Shifted the NCP compute servers hardware on Openstack Based Private Cloud, for efficient utilization of idle resources. Ceph based block storage model is also adopted for flexibility. Currently two computing projects are successfully running on cloud. WLCG T2 Site ( compute Nodes) Local HPC cluster ( Compute Nodes)
12 Network Status Currently 1Gbps of R&E link, between Pakistan Education and Research Network (PERN) and TEIN. Pakistan s Connectivity to TEIN via Singapore (SG) PoP. R&E Link is shared among Universities/Projects in Pakistan ( i.e. Not dedicated for LHC traffic only)
13 T2_PK_NCP Network Status NCP has 1Gbps of Network connectivity with PERN, with following traffic policies: Commercial/commodity Internet 50 Mbps Bandwidth for International R&E destinations is kept un-restricted. Commercial Internet PERN AS45773 Internet2 R&E Networks TEIN GEANT 1Gbps Fully operational on IPv4 and IPv6. R&D link ( without restriction) 1Gbps Commercial IPv6 enabled NCP services ( , DNS, WWW, FTP WLCG Compute Elements (CEs), Storage Element(SE) & Pool nodes, PerfSonar Node 2400:FC00:8540::/ / /27 NCP WLCG site Hu2wai-NE-40
14 Network Traffic Statistics ( 1 Year) ~ 50 % of NCP traffic is on IPv6 ( Mostly WLCG traffic) IPv6 Traffic trend is increasing.
15 T2_PK_NCP Network Status Low Network Throughput is being observed from T1 sites, for last 1 year, which resulted in decommissioning T2_PK_NCP links. To become a commissioned site for production data transfers, a site must pass a commissioning test. [1] T1 to T2 site average data transfer rate must be > 20MB/s. Accumulated transfer rate of all T1 sites T2 to T1 site average data transfer rate must be > 5MB/s. However, NCP is getting average ~ 5MB/s to 10MB/s transfer rate from almost all T1 sites in multiple tests. [1] Accumulated transfer rate of T1_IT_CNAF site
16 T2_PK_NCP Network Status Only T1_UK_RAL -> NCP link commissioned in Nov 2017 and later on decommissioned in April, 2018 due to low data rate.
17 T2_PK_NCP : Site Readiness Report Recently changed status from WR to Morgue, Due to Network issues. Ref:
18 Network issues Multiple simultaneous Network related issues identified, which increased the complexity of issue: T1_US_FNAL, and T1_RU_JINR were not initially following R&E (TEIN) path.( Resolved: April 2018) Low buffer related errors observed in NCP Gateway Device. Replaced with new hardware. (Resolved: March, 2018) PERN also Identified the issue in their uplink route. ( Resolved: March, 2018).
19 PERN Network Complexity NCP is located on North Region, while TEIN link is terminated on south Region HEC(KHI) PoP. NCP Multiple Physical routes and devices exist between the path from NCP to TEIN. Initially PERN network suffered with frequent congestion issues to fiber ring outages. (However now situation is stable, due to bifurcation of fiber routes) TEIN Link Ref:
20 Network issues Steps taken up-till now, for identification of Low throughput from WLCG sites: Involved WLCG Network Throughput Working Group, and Global NOC (IU). Tuning of storage servers: system parameters /etc/sysctl.conf file. Placement of Storage node at NCP edge network. Deployed PerfSonar Nodes at NCP and PERN Network. Installed GridFTP clients on PERN edge to isolate the issue.
21 Network Map for Troubleshooting Network Team from Global NoC and ESNET prepared Map based on Perfsonar nodes in path, Initially investigating low throughput issue from Italian site T1_IT_CNAF, and T1_UK_RAL to NCP Case: 9t4PpYPzZOfJUAhupRAodhPT6HNIZi6J8ljpBtw /edit# Figure: Perfsonar Nodes & Latency between paths
22 PerfSonar Mesh Ref:
23 Findings & conclusion Traffic path from R&E network is offering high latency as compared to commercial network. Difference of 100+ ms. Commercial route is providing better WLCG transfer rate, but have some issues: Can not ensure guaranteed bandwidth all the time either from PERN or far end T1 Site. NCP is currently the major user of 1Gbps R&E link in Pakistan, but ~10-30 % utilization by other organizations, which eventually reducing bandwidth for LHC traffic. Minor packet loss on high latency route can severely impact throughput. Need to increase the coordination among intermediates Network Operators to find out the source of packet loss, possibly low buffered intermediate device, or source of intermittent congestion.
24 Upcoming Work Planning to pledge at least 3 x times more compute and storage to fulfill the needs of WLCG experiment during HL-LHC. Compute pledges from 6K ( HEPSPEC) -> 18 K (HEPSPEC) Storage Pledges from 500 TB -> 2PB. International R&E Bandwidth increase from 1Gbps to 2.5Gbps. Upgradation of storage and ToR switches 1Gbps->10Gbps HL-LHC---Expected Storage Requirement Upgradation of PERN2 -> PERN3 Edge network upgradation to 10Gbps Core Network upgradation to support 40Gbps NCP site connectivity with LHCONE. HL-LHC----Expected Computing Requirement
25 Q &A
t and Migration of WLCG TEIR-2
IPv6 Deploymen t and Migration of WLCG TEIR-2 site resources on Privat te Cloud Saqib Haleem Saqib Haleem National Centre for Physics, Pakistan Agenda Introduction of National Centre for Physics (NCP)
More informationConstant monitoring of multi-site network connectivity at the Tokyo Tier2 center
Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center, T. Mashimo, N. Matsui, H. Matsunaga, H. Sakamoto, I. Ueda International Center for Elementary Particle Physics, The University
More informationGrid Computing Activities at KIT
Grid Computing Activities at KIT Meeting between NCP and KIT, 21.09.2015 Manuel Giffels Karlsruhe Institute of Technology Institute of Experimental Nuclear Physics & Steinbuch Center for Computing Courtesy
More informationSystem upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.
System upgrade and future perspective for the operation of Tokyo Tier2 center, T. Mashimo, N. Matsui, H. Sakamoto and I. Ueda International Center for Elementary Particle Physics, The University of Tokyo
More informationConference The Data Challenges of the LHC. Reda Tafirout, TRIUMF
Conference 2017 The Data Challenges of the LHC Reda Tafirout, TRIUMF Outline LHC Science goals, tools and data Worldwide LHC Computing Grid Collaboration & Scale Key challenges Networking ATLAS experiment
More informationWLCG Network Throughput WG
WLCG Network Throughput WG Shawn McKee, Marian Babik for the Working Group HEPiX Tsukuba 16-20 October 2017 Working Group WLCG Network Throughput WG formed in the fall of 2014 within the scope of WLCG
More informationSupport for multiple virtual organizations in the Romanian LCG Federation
INCDTIM-CJ, Cluj-Napoca, 25-27.10.2012 Support for multiple virtual organizations in the Romanian LCG Federation M. Dulea, S. Constantinescu, M. Ciubancan Department of Computational Physics and Information
More informationAustrian Federated WLCG Tier-2
Austrian Federated WLCG Tier-2 Peter Oettl on behalf of Peter Oettl 1, Gregor Mair 1, Katharina Nimeth 1, Wolfgang Jais 1, Reinhard Bischof 2, Dietrich Liko 3, Gerhard Walzel 3 and Natascha Hörmann 3 1
More informationThe creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM
The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM Lukas Nellen ICN-UNAM lukas@nucleares.unam.mx 3rd BigData BigNetworks Conference Puerto Vallarta April 23, 2015 Who Am I? ALICE
More informationScientific data processing at global scale The LHC Computing Grid. fabio hernandez
Scientific data processing at global scale The LHC Computing Grid Chengdu (China), July 5th 2011 Who I am 2 Computing science background Working in the field of computing for high-energy physics since
More informationHigh-density Grid storage system optimization at ASGC. Shu-Ting Liao ASGC Operation team ISGC 2011
High-density Grid storage system optimization at ASGC Shu-Ting Liao ASGC Operation team ISGC 211 Outline Introduction to ASGC Grid storage system Storage status and issues in 21 Storage optimization Summary
More informationSpanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"
Spanish Tier-2 Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) Introduction Report here the status of the federated T2 for CMS basically corresponding to the budget 2006-2007 concentrate on last year
More informationStatus of KISTI Tier2 Center for ALICE
APCTP 2009 LHC Physics Workshop at Korea Status of KISTI Tier2 Center for ALICE August 27, 2009 Soonwook Hwang KISTI e-science Division 1 Outline ALICE Computing Model KISTI ALICE Tier2 Center Future Plan
More informationClouds at other sites T2-type computing
Clouds at other sites T2-type computing Randall Sobie University of Victoria Randall Sobie IPP/Victoria 1 Overview Clouds are used in a variety of ways for Tier-2 type computing MC simulation, production
More informationDisclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme
NET1343BU NSX Performance Samuel Kommu #VMworld #NET1343BU Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no
More informationFigure 1: cstcdie Grid Site architecture
AccessionIndex: TCD-SCSS-T.20121208.098 Accession Date: Accession By: Object name: cstcdie Grid Site Beowulf Clusters and Datastore Vintage: c.2009 Synopsis: Complex of clusters & storage (1500 cores/600
More informationClouds in High Energy Physics
Clouds in High Energy Physics Randall Sobie University of Victoria Randall Sobie IPP/Victoria 1 Overview Clouds are integral part of our HEP computing infrastructure Primarily Infrastructure-as-a-Service
More informationData transfer over the wide area network with a large round trip time
Journal of Physics: Conference Series Data transfer over the wide area network with a large round trip time To cite this article: H Matsunaga et al 1 J. Phys.: Conf. Ser. 219 656 Recent citations - A two
More informationGrid Computing at the IIHE
BNC 2016 Grid Computing at the IIHE The Interuniversity Institute for High Energies S. Amary, F. Blekman, A. Boukil, O. Devroede, S. Gérard, A. Ouchene, R. Rougny, S. Rugovac, P. Vanlaer, R. Vandenbroucke
More informationDistributing storage of LHC data - in the nordic countries
Distributing storage of LHC data - in the nordic countries Gerd Behrmann INTEGRATE ASG Lund, May 11th, 2016 Agenda WLCG: A world wide computing grid for the LHC NDGF: The Nordic Tier 1 dcache: Distributed
More informationNetwork Analytics. Hendrik Borras, Marian Babik IT-CM-MM
Network Analytics Hendrik Borras, Marian Babik IT-CM-MM perfsonar Infrastructure perfsonar has been widely deployed in WLCG 249 active instances, deployed at 120 sites including major network hubs at ESNet,
More informationCLOUDS OF JINR, UNIVERSITY OF SOFIA AND INRNE JOIN TOGETHER
CLOUDS OF JINR, UNIVERSITY OF SOFIA AND INRNE JOIN TOGETHER V.V. Korenkov 1, N.A. Kutovskiy 1, N.A. Balashov 1, V.T. Dimitrov 2,a, R.D. Hristova 2, K.T. Kouzmov 2, S.T. Hristov 3 1 Laboratory of Information
More informationOverview. About CERN 2 / 11
Overview CERN wanted to upgrade the data monitoring system of one of its Large Hadron Collider experiments called ALICE (A La rge Ion Collider Experiment) to ensure the experiment s high efficiency. They
More informationNew International Connectivities of SINET5
Mar 28 th, 2018 at APAN45 New International Connectivities of SINET5 SINET 100G Global Ring Motonori Nakamura National Institute of Informatics (NII), Japan Academic Infrastructure operated by NII Our
More informationPresentation of the LHCONE Architecture document
Presentation of the LHCONE Architecture document Marco Marletta, GARR LHCONE Meeting Paris, Tuesday 5th April 2011 Agenda Background Design Definitions Architecture Services Policy Next steps 2 Background
More informationLHCb Computing Resources: 2018 requests and preview of 2019 requests
LHCb Computing Resources: 2018 requests and preview of 2019 requests LHCb-PUB-2017-009 23/02/2017 LHCb Public Note Issue: 0 Revision: 0 Reference: LHCb-PUB-2017-009 Created: 23 rd February 2017 Last modified:
More informationConnectivity Services, Autobahn and New Services
Connectivity Services, Autobahn and New Services Domenico Vicinanza, DANTE EGEE 09, Barcelona, 21 st -25 th September 2009 Agenda Background GÉANT Connectivity services: GÉANT IP GÉANT Plus GÉANT Lambda
More informationMcAfee Network Security Platform 9.2
McAfee Network Security Platform 9.2 (9.2.7.22-9.2.7.20 Manager-Virtual IPS Release Notes) Contents About this release New features Enhancements Resolved issues Installation instructions Known issues Product
More informationHigh-Energy Physics Data-Storage Challenges
High-Energy Physics Data-Storage Challenges Richard P. Mount SLAC SC2003 Experimental HENP Understanding the quantum world requires: Repeated measurement billions of collisions Large (500 2000 physicist)
More informationThe INFN Tier1. 1. INFN-CNAF, Italy
IV WORKSHOP ITALIANO SULLA FISICA DI ATLAS E CMS BOLOGNA, 23-25/11/2006 The INFN Tier1 L. dell Agnello 1), D. Bonacorsi 1), A. Chierici 1), M. Donatelli 1), A. Italiano 1), G. Lo Re 1), B. Martelli 1),
More informationAndrea Sciabà CERN, Switzerland
Frascati Physics Series Vol. VVVVVV (xxxx), pp. 000-000 XX Conference Location, Date-start - Date-end, Year THE LHC COMPUTING GRID Andrea Sciabà CERN, Switzerland Abstract The LHC experiments will start
More informationLong Term Data Preservation for CDF at INFN-CNAF
Long Term Data Preservation for CDF at INFN-CNAF S. Amerio 1, L. Chiarelli 2, L. dell Agnello 3, D. De Girolamo 3, D. Gregori 3, M. Pezzi 3, A. Prosperini 3, P. Ricci 3, F. Rosso 3, and S. Zani 3 1 University
More informationResearch and Education Networking Ecosystem and NSRC Assistance
Research and Education Networking Ecosystem and NSRC Assistance This document is a result of work by the Network Startup Resource Center (NSRC at http://www.nsrc.org). This document may be freely copied,
More informationThe Legnaro-Padova distributed Tier-2: challenges and results
The Legnaro-Padova distributed Tier-2: challenges and results Simone Badoer a, Massimo Biasotto a,fulviacosta b, Alberto Crescente b, Sergio Fantinel a, Roberto Ferrari b, Michele Gulmini a, Gaetano Maron
More informationBlueGene/L. Computer Science, University of Warwick. Source: IBM
BlueGene/L Source: IBM 1 BlueGene/L networking BlueGene system employs various network types. Central is the torus interconnection network: 3D torus with wrap-around. Each node connects to six neighbours
More informationGrid and Cloud Activities in KISTI
Grid and Cloud Activities in KISTI March 23, 2011 Soonwook Hwang KISTI, KOREA 1 Outline Grid Operation and Infrastructure KISTI ALICE Tier2 Center FKPPL VO: Production Grid Infrastructure Global Science
More informationAnalytics of Wide-Area Lustre Throughput Using LNet Routers
Analytics of Wide-Area Throughput Using LNet Routers Nagi Rao, Neena Imam, Jesse Hanley, Sarp Oral Oak Ridge National Laboratory User Group Conference LUG 2018 April 24-26, 2018 Argonne National Laboratory
More informationCMS Data Transfer Challenges and Experiences with 40G End Hosts
CMS Data Transfer Challenges and Experiences with 40G End Hosts NEXT Technology Exchange Advanced Networking / Joint Techs Indianapolis, October 2014 Azher Mughal, Dorian Kcira California Institute of
More informationGÉANT Mission and Services
GÉANT Mission and Services Vincenzo Capone Senior Technical Business Development Officer CREMLIN WP2 Workshop on Big Data Management 15 February 2017, Moscow GÉANT - Networks Manages research & education
More informationMicrosoft IT Leverages its Compute Service to Virtualize SharePoint 2010
Microsoft IT Leverages its Compute Service to Virtualize SharePoint 2010 Published: June 2011 The following content may no longer reflect Microsoft s current position or infrastructure. This content should
More informationvstart 50 VMware vsphere Solution Specification
vstart 50 VMware vsphere Solution Specification Release 1.3 for 12 th Generation Servers Dell Virtualization Solutions Engineering Revision: A00 March 2012 THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES
More informationOpportunities A Realistic Study of Costs Associated
e-fiscal Summer Workshop Opportunities A Realistic Study of Costs Associated X to Datacenter Installation and Operation in a Research Institute can we do EVEN better? Samos, 3rd July 2012 Jesús Marco de
More informationOperating the Distributed NDGF Tier-1
Operating the Distributed NDGF Tier-1 Michael Grønager Technical Coordinator, NDGF International Symposium on Grid Computing 08 Taipei, April 10th 2008 Talk Outline What is NDGF? Why a distributed Tier-1?
More informationData storage services at KEK/CRC -- status and plan
Data storage services at KEK/CRC -- status and plan KEK/CRC Hiroyuki Matsunaga Most of the slides are prepared by Koichi Murakami and Go Iwai KEKCC System Overview KEKCC (Central Computing System) The
More informationATLAS 実験コンピューティングの現状と将来 - エクサバイトへの挑戦 坂本宏 東大 ICEPP
ATLAS 実験コンピューティングの現状と将来 - エクサバイトへの挑戦 坂本宏 東大 ICEPP 1 Contents Energy Frontier Particle Physics Large Hadron Collider (LHC) LHC Experiments: mainly ATLAS Requirements on computing Worldwide LHC Computing
More informationAMRES Combining national, regional and & EU efforts
AMRES Combining national, regional and & EU efforts Eastern Europe Partnership Event, 8 November 2011, Bucharest www.amres.ac.rs 1 www.amres.ac.rs Who are we? - Why AMRES/UoB? In 90s UoB (computer centre)
More informationParallel Storage Systems for Large-Scale Machines
Parallel Storage Systems for Large-Scale Machines Doctoral Showcase Christos FILIPPIDIS (cfjs@outlook.com) Department of Informatics and Telecommunications, National and Kapodistrian University of Athens
More informationMeasurement and Monitoring. Yasuichi Kitmaura Takatoshi Ikeda
Measurement and Monitoring Yasuichi Kitmaura (kita@jp.apan.net) Takatoshi Ikeda (ikeda@kddnet.ad.jp) Who are we? Yasuichi Kitamura researcher of National Institute of Information and Communications Technology
More informationCloud Computing For Researchers
Cloud Computing For Researchers August, 2016 Compute Canada is often asked about the potential of outsourcing to commercial clouds. This has been investigated as an alternative or supplement to purchasing
More informationNetwork Expectations for Data Intensive Science
Network Expectations for Data Intensive Science Introduction This guide is aimed at researchers who are running or planning to run data intensive science applications over the Janet network. As it becomes
More informationUsers and utilization of CERIT-SC infrastructure
Users and utilization of CERIT-SC infrastructure Equipment CERIT-SC is an integral part of the national e-infrastructure operated by CESNET, and it leverages many of its services (e.g. management of user
More informationServer Specifications
Requirements Server s It is highly recommended that MS Exchange does not run on the same server as Practice Evolve. Server Minimum Minimum spec. is influenced by choice of operating system and by number
More informationTowards Network Awareness in LHC Computing
Towards Network Awareness in LHC Computing CMS ALICE CERN Atlas LHCb LHC Run1: Discovery of a New Boson LHC Run2: Beyond the Standard Model Gateway to a New Era Artur Barczyk / Caltech Internet2 Technology
More informationI Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011
I Tier-3 di CMS-Italia: stato e prospettive Claudio Grandi Workshop CCR GRID 2011 Outline INFN Perugia Tier-3 R&D Computing centre: activities, storage and batch system CMS services: bottlenecks and workarounds
More informationDYNES: DYnamic NEtwork System
DYNES: DYnamic NEtwork System Artur Barczyk California Institute of Technology / US LHCNet TERENA e2e Workshop Prague, November 29 th, 2010 1 OUTLINE What is DYNES Status Deployment Plan 2 DYNES Overview
More informationPaperspace. Architecture Overview. 20 Jay St. Suite 312 Brooklyn, NY Technical Whitepaper
Architecture Overview Copyright 2016 Paperspace, Co. All Rights Reserved June - 1-2017 Technical Whitepaper Paperspace Whitepaper: Architecture Overview Content 1. Overview 3 2. Virtualization 3 Xen Hypervisor
More informationProfiling Grid Data Transfer Protocols and Servers. George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA
Profiling Grid Data Transfer Protocols and Servers George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA Motivation Scientific experiments are generating large amounts of data Education
More informationStorage and I/O requirements of the LHC experiments
Storage and I/O requirements of the LHC experiments Sverre Jarp CERN openlab, IT Dept where the Web was born 22 June 2006 OpenFabrics Workshop, Paris 1 Briefly about CERN 22 June 2006 OpenFabrics Workshop,
More informationIvane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU
Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU Grid cluster at the Institute of High Energy Physics of TSU Authors: Arnold Shakhbatyan Prof. Zurab Modebadze Co-authors:
More informationAccelerating Throughput from the LHC to the World
Accelerating Throughput from the LHC to the World David Groep David Groep Nikhef PDP Advanced Computing for Research v5 Ignatius 2017 12.5 MByte/event 120 TByte/s and now what? Kans Higgs deeltje: 1 op
More informationThe Software Defined Online Storage System at the GridKa WLCG Tier-1 Center
The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center CHEP 2018, Sofia Jan Erik Sundermann, Jolanta Bubeliene, Ludmilla Obholz, Andreas Petzold STEINBUCH CENTRE FOR COMPUTING (SCC)
More informationFUJITSU Software Smart Communication Optimizer V1.1.0 Introduction
FUJITSU Software V1.1.0 Introduction December, 2018 Fujitsu Limited Contents Introduction What is? Features Application examples Operating environment and System configuration Product information 1 Introduction
More informationHigh Throughput WAN Data Transfer with Hadoop-based Storage
High Throughput WAN Data Transfer with Hadoop-based Storage A Amin 2, B Bockelman 4, J Letts 1, T Levshina 3, T Martin 1, H Pi 1, I Sfiligoi 1, M Thomas 2, F Wuerthwein 1 1 University of California, San
More informationCurrent Status of the Ceph Based Storage Systems at the RACF
Journal of Physics: Conference Series PAPER OPEN ACCESS Current Status of the Ceph Based Storage Systems at the RACF To cite this article: A. Zaytsev et al 2015 J. Phys.: Conf. Ser. 664 042027 View the
More information70-414: Implementing an Advanced Server Infrastructure Course 01 - Creating the Virtualization Infrastructure
70-414: Implementing an Advanced Server Infrastructure Course 01 - Creating the Virtualization Infrastructure Slide 1 Creating the Virtualization Infrastructure Slide 2 Introducing Microsoft System Center
More informationComputing for LHC in Germany
1 Computing for LHC in Germany Günter Quast Universität Karlsruhe (TH) Meeting with RECFA Berlin, October 5th 2007 WLCG Tier1 & Tier2 Additional resources for data analysis - HGF ''Physics at the Terascale''
More informationReference Architecture for Dell VIS Self-Service Creator and VMware vsphere 4
Reference Architecture for Dell VIS Self-Service Creator and VMware vsphere 4 Solutions for Small & Medium Environments Virtualization Solutions Engineering Ryan Weldon and Tom Harrington THIS WHITE PAPER
More informationGRID AND HPC SUPPORT FOR NATIONAL PARTICIPATION IN LARGE-SCALE COLLABORATIONS
GRID AND HPC SUPPORT FOR NATIONAL PARTICIPATION IN LARGE-SCALE COLLABORATIONS M. Dulea, S. Constantinescu, M. Ciubancan, T. Ivanoaica, C. Placinta, I.T. Vasile, D. Ciobanu-Zabet Department of Computational
More informationBenoit DELAUNAY Benoit DELAUNAY 1
Benoit DELAUNAY 20091023 Benoit DELAUNAY 1 CC-IN2P3 provides computing and storage for the 4 LHC experiments and many others (astro particles...) A long history of service sharing between experiments Some
More informationWorldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010
Worldwide Production Distributed Data Management at the LHC Brian Bockelman MSST 2010, 4 May 2010 At the LHC http://op-webtools.web.cern.ch/opwebtools/vistar/vistars.php?usr=lhc1 Gratuitous detector pictures:
More informationCMS Belgian T2. G. Bruno UCL, Louvain, Belgium on behalf of the CMS Belgian T2 community. GridKa T1/2 meeting, Karlsruhe Germany February
UCL, Louvain, Belgium on behalf of the community GridKa, Karlsruhe Germany February 2 2006 community Institutes in CMS (40 physicists) UA (Antwerp) UCL (Louvain-La-Neuve) ULB (Brussels) UMH (Mons) VUB
More informationANSE: Advanced Network Services for [LHC] Experiments
ANSE: Advanced Network Services for [LHC] Experiments Artur Barczyk California Institute of Technology Joint Techs 2013 Honolulu, January 16, 2013 Introduction ANSE is a project funded by NSF s CC-NIE
More informationUse of containerisation as an alternative to full virtualisation in grid environments.
Journal of Physics: Conference Series PAPER OPEN ACCESS Use of containerisation as an alternative to full virtualisation in grid environments. Related content - Use of containerisation as an alternative
More informationarxiv: v1 [cs.dc] 20 Jul 2015
Designing Computing System Architecture and Models for the HL-LHC era arxiv:1507.07430v1 [cs.dc] 20 Jul 2015 L Bauerdick 1, B Bockelman 2, P Elmer 3, S Gowdy 1, M Tadel 4 and F Würthwein 4 1 Fermilab,
More informationPreparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch
Preparing for High-Luminosity LHC Bob Jones CERN Bob.Jones cern.ch The Mission of CERN Push back the frontiers of knowledge E.g. the secrets of the Big Bang what was the matter like within the first
More informationVirtualization. A very short summary by Owen Synge
Virtualization A very short summary by Owen Synge Outline What is Virtulization? What's virtulization good for? What's virtualisation bad for? We had a workshop. What was presented? What did we do with
More informationDatabase Services at CERN with Oracle 10g RAC and ASM on Commodity HW
Database Services at CERN with Oracle 10g RAC and ASM on Commodity HW UKOUG RAC SIG Meeting London, October 24 th, 2006 Luca Canali, CERN IT CH-1211 LCGenève 23 Outline Oracle at CERN Architecture of CERN
More informationNetworking for Data Acquisition Systems. Fabrice Le Goff - 14/02/ ISOTDAQ
Networking for Data Acquisition Systems Fabrice Le Goff - 14/02/2018 - ISOTDAQ Outline Generalities The OSI Model Ethernet and Local Area Networks IP and Routing TCP, UDP and Transport Efficiency Networking
More informationBest Practices for Deploying a Mixed 1Gb/10Gb Ethernet SAN using Dell Storage PS Series Arrays
Best Practices for Deploying a Mixed 1Gb/10Gb Ethernet SAN using Dell Storage PS Series Arrays Dell EMC Engineering December 2016 A Dell Best Practices Guide Revisions Date March 2011 Description Initial
More informationCMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster
CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster Vaikunth Thukral Department of Physics and Astronomy Texas A&M University 1 Outline Grid Computing with CMS:
More informationGÉANT : e-infrastructure connectivity for the data deluge
GÉANT : e-infrastructure connectivity for the data deluge Richard Hughes-Jones, Maria Minaricova & Vincenzo Capone DANTE BioMedBridges Workshop, Hinxton, 15-16 May 2014 What is GÉANT? High-bandwidth, high-performance
More informationG-NET: Effective GPU Sharing In NFV Systems
G-NET: Effective Sharing In NFV Systems Kai Zhang*, Bingsheng He^, Jiayu Hu #, Zeke Wang^, Bei Hua #, Jiayi Meng #, Lishan Yang # *Fudan University ^National University of Singapore #University of Science
More informationSLIDE 1 - COPYRIGHT 2015 ELEPHANT FLOWS IN THE ROOM: SCIENCEDMZ NATIONALLY DISTRIBUTED
SLIDE 1 - COPYRIGHT 2015 ELEPHANT FLOWS IN THE ROOM: SCIENCEDMZ NATIONALLY DISTRIBUTED SLIDE 2 - COPYRIGHT 2015 Do you know what your campus network is actually capable of? (i.e. have you addressed your
More informationAchieving the Science DMZ
Achieving the Science DMZ Eli Dart, Network Engineer ESnet Network Engineering Group Joint Techs, Winter 2012 Baton Rouge, LA January 22, 2012 Outline of the Day Motivation Services Overview Science DMZ
More informationPI SERVER 2012 Do. More. Faster. Now! Copyri g h t 2012 OSIso f t, LLC.
PI SERVER 2012 Do. More. Faster. Now! Copyri g h t 2012 OSIso f t, LLC. AUGUST 7, 2007 APRIL 14, 2010 APRIL 24, 2012 Copyri g h t 2012 OSIso f t, LLC. 2 PI SERVER 2010 PERFORMANCE 2010 R3 Max Point Count
More informationEnhancing Infrastructure: Success Stories
Enhancing Infrastructure: Success Stories Eli Dart, Network Engineer ESnet Network Engineering Group Joint Techs, Winter 2012 Baton Rouge, LA January 24, 2012 Outline Motivation for strategic investments
More informationMcAfee Network Security Platform 9.2
Revision B McAfee Network Security Platform 9.2 (9.2.7.9-9.2.7.10 Manager-Virtual IPS Release Notes) Contents About this release New features Enhancements Resolved issues Installation instructions Known
More informationVMware vcloud Air User's Guide
vcloud Air This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of this document,
More informationNCIT*net 2 General Information
NCIT*net 2 General Information NCIT*net 2 An Introduction A component of NCCT research team activities is the development of a next generation network to promote the interaction of architecture, technologies,
More informationPrivate Cloud at IIT Delhi
Private Cloud at IIT Delhi Success Story Engagement: Long Term Industry: Education Offering: Private Cloud Deployment Business Challenge IIT Delhi, one of the India's leading educational Institute wanted
More informationBig Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback
Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy Texas A&M Big Data Workshop October 2011 January 2015, Texas A&M University Research Topics Seminar 1 Outline Overview of
More informationEurope and its Open Science Cloud: the Italian perspective. Luciano Gaido Plan-E meeting, Poznan, April
Europe and its Open Science Cloud: the Italian perspective Luciano Gaido (gaido@to.infn.it) Plan-E meeting, Poznan, April 27 2017 Background Italy has a long-standing expertise and experience in the management
More informationFlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC
white paper FlashGrid Software Intel SSD DC P3700/P3600/P3500 Topic: Hyper-converged Database/Storage FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC Abstract FlashGrid
More informationLHC and LSST Use Cases
LHC and LSST Use Cases Depots Network 0 100 200 300 A B C Paul Sheldon & Alan Tackett Vanderbilt University LHC Data Movement and Placement n Model must evolve n Was: Hierarchical, strategic pre- placement
More informationALICE Grid Activities in US
ALICE Grid Activities in US 1 ALICE-USA Computing Project ALICE-USA Collaboration formed to focus on the ALICE EMCal project Construction, installation, testing and integration participating institutions
More informationThe LHC Computing Grid
The LHC Computing Grid Visit of Finnish IT Centre for Science CSC Board Members Finland Tuesday 19 th May 2009 Frédéric Hemmer IT Department Head The LHC and Detectors Outline Computing Challenges Current
More informationCloud Service SLA Declaration
Cloud Service SLA Declaration Basic level of support for Cloud services (SLA 1) Table of Content: 1. Definitions 2. General terms 3. Level of service warranty service functioning 4. Provider`s liability
More informationFirst Experience with LCG. Board of Sponsors 3 rd April 2009
First Experience with LCG Operation and the future... CERN openlab Board of Sponsors 3 rd April 2009 Ian Bird LCG Project Leader The LHC Computing Challenge Signal/Noise: 10-9 Data volume High rate * large
More informationMidoNet Scalability Report
MidoNet Scalability Report MidoNet Scalability Report: Virtual Performance Equivalent to Bare Metal 1 MidoNet Scalability Report MidoNet: For virtual performance equivalent to bare metal Abstract: This
More informationEdge for All Business
1 Edge for All Business Datasheet Zynstra is designed and built for the edge the business-critical compute activity that takes place outside a large central datacenter, in branches, remote offices, or
More information