Grid Computing Activities at KIT Meeting between NCP and KIT, 21.09.2015 Manuel Giffels Karlsruhe Institute of Technology Institute of Experimental Nuclear Physics & Steinbuch Center for Computing Courtesy Argonne National Laboratory 1 KIT University of the State of Baden-Württemberg and National Research Center of the Helmholtz Association www.kit.edu
Grid Computing Today Tier 0 7(8) Tier 1 ~50 Tier 2 Tier 3 2 More than 50 CMS centers, in more then 20 countries Flags taken from Wikipedia: http://de.wikipedia.org/wiki/liste_der_nationalflaggen Christoph Wissing Tier-2 Candidates: - Thailand - Malaysia > Tier 0 W-LCG Resources: prompt reconstruction 590k logical CPU cores 325,000 Terabyte Disk store RAW data and 257,000 export to T1s Terabyte Tape > Tier 1 2 Million jobs/day re-recontruction 20 GB/s network transfers long term storage of RAW and MC data German CMS Contribution: > Tier 2 GridKa TIER 1 MC production ~ 10 % (2015) User analysis Aachen/Hamburg TIER 2 > Tier 3 ~ 8 % (2015) Mainly user analysis World-LHC Computing Grid (W-LCG) Rebus
GridKa TIER 1 3
GridKa History Established 2002 on request of the German HEP community First customers during the startup phase BABAR, D0, CDF Currently supporting ALICE, ATLAS, CMS, LHCb BABAR, Belle (2), Compass Auger 4
GridKa Farm Batch System: Univa Grid Engine ~ 630 Worker Nodes, ~ 16300 logical cores, ~ 26 TB RAM ~ 13000 job slots for single and multi-core jobs Storage System: dcache ~ 10,000 TB disk storage ~ 46,000 TB tape capacity Network Connectivity 100 Gbit/s connectivity to LHCOPN (WLCG Tier0-Tier1 Private Network) 100 Gbit/s connectivity to LHCONE (LHC Open Network Environment) Very stable operation! 5
GridKa T1 Resources for CMS Pledged Resources: CPU: 1382 job slots, 26850 HEPSpec2006 2600 TB disk space 7400 TB tape space Opportunistic resource allocation is allowed Actual CPU usage is normally over the pledge Resources for the national CMS community: CPU: Opportunistic resource usage Separate storage instance shared with ATLAS 340 TB total disk space, 170 TB allocated to CMS Access to tape possible Support for the national analysis communities (Aachen, Hamburg, Karlsruhe) 6
GridKa Contribution to CMS 12 % of CMS data (both MC and recorded data) stored at GridKa Around 10% of CMS TIER 1 jobs running at GridKa 7
CMS GridKa TIER 1 Support Team Expert rotation scheme Shift leader: 3 senior group members taking shift leadership for 4 to 6 weeks Shifter: 4 PhD students taking shifts for 1 week each Support effort is credited as CMS service work 2.6 FTE per year dedicated to CMS GridKa TIER 1 support and development 8
CMS Support Team Operations Regular monitoring of CMS operations at GridKa TIER 1 Data transfer and storage management Management of CMS site specific services Troubleshooting and issue follow-up: Responding to and opening tickets Mailing lists monitoring Attend and report to weekly coordination meetings CMS site support and computing operations Institute computing meeting GridKa TIER 1 Middleware meeting 9
Education Program Annual International GridKa School on advanced computing technologies organized by GridKa TIER 1 center 50% Plenary presentations 50% Hands-on courses 2014:120 participants / 19 countries Audience: Grid and cloud newbies Advanced users Administrators Graduate and PhD students Also participants from industry 10
Development Projects Much more computing activities are ongoing at IEKP: Development of CMS data transfer and data management tools High-throughput data analysis cluster - based on local SSD caching Virtualisation and cloud computing - institute desktop cloud, Freiburg HPC Cloud, planned cooperation with commercial cloud service providers HappyFace (meta-monitoring tool) development Data Requests Cache Drives (SSD/HD) Physicists Cache Scheduler Worker Cache Mapping Network file space 11
Questions? 12