Flying HTCondor at 100gbps Over the Golden State

Similar documents
Shooting for the sky: Testing the limits of condor. HTCondor Week May 2015 Edgar Fajardo On behalf of OSG Software and Technology

glideinwms architecture by Igor Sfiligoi, Jeff Dost (UCSD)

PRP. Frank Würthwein (SDSC/UCSD) Jason Nielsen (UCSC) Owen Long (UCR) Chris West (UCSB) Anyes Taffard (UCI) Maria Spiropolu (Caltech)

A Virtual Comet. HTCondor Week 2017 May Edgar Fajardo On behalf of OSG Software and Technology

Introducing the HTCondor-CE

glideinwms Training Glidein Internals How they work and why by Igor Sfiligoi, Jeff Dost (UCSD) glideinwms Training Glidein internals 1

Condor-G: HTCondor for grid submission. Jaime Frey (UW-Madison), Jeff Dost (UCSD)

One Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool. D. Mason for CMS Software & Computing

Look What I Can Do: Unorthodox Uses of HTCondor in the Open Science Grid

Building Campus HTC Sharing Infrastructures. Derek Weitzel University of Nebraska Lincoln (Open Science Grid Hat)

glideinwms: Quick Facts

The HTCondor CacheD. Derek Weitzel, Brian Bockelman University of Nebraska Lincoln

glideinwms UCSD Condor tunning by Igor Sfiligoi (UCSD) UCSD Jan 18th 2012 Condor Tunning 1

Scheduling Computational and Storage Resources on the NRP

Configuring a glideinwms factory

What s new in HTCondor? What s coming? HTCondor Week 2018 Madison, WI -- May 22, 2018

Care and Feeding of HTCondor Cluster. Steven Timm European HTCondor Site Admins Meeting 8 December 2014

Factory Ops Site Debugging

On-demand provisioning of HEP compute resources on cloud sites and shared HPC centers

Primer for Site Debugging

Networking and High Throughput Computing. Garhan Attebury HTCondor Week 2015

Singularity in CMS. Over a million containers served

An update on the scalability limits of the Condor batch system

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

VC3. Virtual Clusters for Community Computation. DOE NGNS PI Meeting September 27-28, 2017

Eurogrid: a glideinwms based portal for CDF data analysis - 19th January 2012 S. Amerio. (INFN Padova) on behalf of Eurogrid support group

glideinwms Frontend Installation

UW-ATLAS Experiences with Condor

ATLAS Experiment and GCE

CHTC Policy and Configuration. Greg Thain HTCondor Week 2017

Connecting Restricted, High-Availability, or Low-Latency Resources to a Seamless Global Pool for CMS

HTCondor overview. by Igor Sfiligoi, Jeff Dost (UCSD)

Scientific Computing on Emerging Infrastructures. using HTCondor

ALICE Grid Activities in US

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

High Throughput WAN Data Transfer with Hadoop-based Storage

PROOF-Condor integration for ATLAS

Enabling Distributed Scientific Computing on the Campus

Using Puppet to contextualize computing resources for ATLAS analysis on Google Compute Engine

HIGH ENERGY PHYSICS ON THE OSG. Brian Bockelman CCL Workshop, 2016

WLCG Lightweight Sites

Scientific Cluster Deployment and Recovery Using puppet to simplify cluster management

The Echo Project An Update. Alastair Dewhurst, Alison Packer, George Vasilakakos, Bruno Canning, Ian Johnson, Tom Byrne

CERN: LSF and HTCondor Batch Services

California State University (CSU) Fresno Tier-3

Distributed production managers meeting. Armando Fella on behalf of Italian distributed computing group

LHCb experience running jobs in virtual machines

Interfacing HTCondor-CE with OpenStack: technical questions

Grid Computing Activities at KIT

An overview of batch processing. 1-June-2017

UK Tier-2 site evolution for ATLAS. Alastair Dewhurst

HTCondor Week 2015: Implementing an HTCondor service at CERN

HTCondor on Titan. Wisconsin IceCube Particle Astrophysics Center. Vladimir Brik. HTCondor Week May 2018

Measuring and Monitoring Our Networks Performance & Reliability And Prototyping Affordable Cloud Storage/Compute

Pegasus Workflow Management System. Gideon Juve. USC Informa3on Sciences Ins3tute

Data Transfers Between LHC Grid Sites Dorian Kcira

Cloud Computing. Summary

Distributed Caching Using the HTCondor CacheD

Day 9: Introduction to CHTC

Pegasus WMS Automated Data Management in Shared and Nonshared Environments

SLATE. Services Layer at the Edge. First Meeting of the National Research Platform Montana State University August 7-8, 2017

Introduction to Distributed HTC and overlay systems

OSG Lessons Learned and Best Practices. Steven Timm, Fermilab OSG Consortium August 21, 2006 Site and Fabric Parallel Session

Improvements to Configuration. John (TJ) Knoeller Condor Week 2014

UGP and the UC Grid Portals

Austrian Federated WLCG Tier-2

CMS experience of running glideinwms in High Availability mode

BOSCO Architecture. Derek Weitzel University of Nebraska Lincoln

CCB The Condor Connection Broker. Dan Bradley Condor Project CS and Physics Departments University of Wisconsin-Madison

Open Science Grid LATBauerdick/Fermilab

BOSCO Architecture. Derek Weitzel University of Nebraska Lincoln

Things you may not know about HTCondor. John (TJ) Knoeller Condor Week 2017

Autonomic Condor Clouds. David Wolinsky ACIS P2P Group University of Florida

Evolution of Cloud Computing in ATLAS

The Pacific Research Platform (PRP)

Managing large-scale workflows with Pegasus

Monitoring and Analytics With HTCondor Data

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

Evolution of the HEP Content Distribution Network. Dave Dykstra CernVM Workshop 6 June 2016

Running HEP Workloads on Distributed Clouds

Analisi Tier2 e Tier3 Esperienze ai Tier-2 Giacinto Donvito INFN-BARI

Cornell Red Cloud: Campus-based Hybrid Cloud. Steven Lee Cornell University Center for Advanced Computing

Monitoring Primer HTCondor Week 2017 Todd Tannenbaum Center for High Throughput Computing University of Wisconsin-Madison

Outline. ASP 2012 Grid School

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

The ATLAS Distributed Analysis System

Containerized Cloud Scheduling Environment

New Directions and BNL

Things you may not know about HTCondor. John (TJ) Knoeller Condor Week 2017

The Legnaro-Padova distributed Tier-2: challenges and results

Implementation of the Pacific Research Platform over Pacific Wave

DataLink Learn (SaaS or 9.1 Oct 2014+) Integration

Tutorial 4: Condor. John Watt, National e-science Centre

Monitoring the ALICE Grid with MonALISA

Tier 3 batch system data locality via managed caches

Singularity tests. ADC TIM at CERN 21 Sep E. Vamvakopoulos

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

Introduction to HPC Resources and Linux

MONitoring Agents using a Large Integrated Services Architecture. Iosif Legrand California Institute of Technology

Sourcing Director. Setup Internal Supplier View. Setup Internal Supplier View

Transcription:

Flying HTCondor at 100gbps Over the Golden State Jeff Dost (UCSD) HTCondor Week 2016 1

What is PRP? Pacific Research Platform: - 100 gbit network extending from Southern California to Washington - Interconnects Science DMZ between institutions HTCondor Week 2016 2

UW Seattle LHC Scientists across nine West Coast Universities NERSC Compute Resource SLAC Data & Compute Resource UCSC UCD to benefit from Petascale Data & Compute Resources across Pacfific Research Platform CSU Fresno UCSB CALTECH Data & Compute Resource UCI UCR UCSD & UCSC Data & Compute Resource HTCondor Week 2016 3

What is LHC @ UC? Pilot project that unifies University of California Tier 3 campus compute infrastructures on top of PRP network Provides each UC the ability to: - Utilize external compute resources Provided by GlideinWMS HTCondor pool - Access data from anywhere Provided by XRootD Integrates seamlessly with local compute and storage resources HTCondor Week 2016 4

What is LHC @ UC? Participating UCs: ATLAS CMS Resources currently provided: - Each UC - 50k core Comet cluster at SDSC Davis Santa Cruz Eventually: - Any other non-uc participating PRP site - Any OSG site beyond PRP - Other NSF XSEDE and DOE super computing centers - Cloud resources Santa Barbara Riverside Irvine SDSC HTCondor Week 2016 5

Data Access Built on top of XRootD Jobs don t need to run where the data is located Local UC and external compute resources both cache remote data accesses Arbitrary data from local UC can be exported and made available to all compute resources HTCondor Week 2016 6

Hardware shipped to UCs (aka the brick ) Hardware: 40 cores 12 x 4TB data disks 128 GB ram 2 x 10 gbit network interface Software: Full HTCondor pool XRootD server, redirector, and proxy cache cvmfs w/ optional Squid The brick is effectively a site in a box HTCondor Week 2016 7

Traditional Submission universe = vanilla Collector Negotiator Central Manager 3.0 4.0 Submit host Startd Worker Node Startd Worker Node Startd Idle Worker Node Startd Idle Worker Node HTCondor Week 2016 8

Submit to Brick universe = grid grid_resource = condor uclhc-1.ucsc.edu 192.168.x.y x509userproxy = /tmp/x509up_500 +local = true Collector Negotiator 3.0 3.0 Slot 1 Slot 3 Idle 4.0 4.0 Submit host Slot 2 Slot 4 Idle Startd Brick HTCondor Week 2016 9

Submit to UCs universe = grid grid_resource = condor uclhc-1.ucsc.edu 192.168.x.y x509userproxy = /tmp/x509up_500 +uc = true Collector Negotiator Flocking CM @ UCSD 3.0 4.0 3.0 4.0 Collector Negotiator UCD WN Idle UCSC WN Submit host Brick Idle Factory GWMS Magic! UCR WN UCR WN HTCondor Week 2016 10

Submit to Comet universe = grid grid_resource = condor uclhc-1.ucsc.edu 192.168.x.y x509userproxy = /tmp/x509up_500 +sdsc = true Collector Negotiator Flocking CM @ UCSD 3.0 4.0 3.0 4.0 Collector Negotiator Comet WN Idle Comet WN Submit host Brick Idle Factory GWMS Magic! Comet WN Comet WN HTCondor Week 2016 11

site_local to replace vanilla universe = grid grid_resource = condor uclhc-1.ucsc.edu 192.168.x.y x509userproxy = /tmp/x509up_500 +site_local = true Collector Negotiator Flocking CM @ UCSD 3.0 4.0 3.0 4.0 Collector Negotiator UCSC WN Idle UCSC WN Submit host Brick Idle Factory GWMS Magic! UCSC WN UCSC WN HTCondor Week 2016 12

Why site_local vs vanilla? We dynamically set xrootd cache location using startup scripts: # get cache address based on where we land _site=`grep -i "^GLIDEIN_Site " $_config awk '{print $2}'` CMS_XROOTD_CACHE=`grep -i "$_site"_cms client/xrootd-cache-location.txt awk '{print $2}'` ATLAS_XROOTD_CACHE=`grep -i "$_site"_atlas client/xrootd-cache-location.txt awk '{print $2}'` # gwms way to set env vars into user job add_config_line CMS_XROOTD_CACHE "$CMS_XROOTD_CACHE" add_condor_vars_line CMS_XROOTD_CACHE "S" "-" "+" "Y" "Y" "+" add_config_line ATLAS_XROOTD_CACHE "$ATLAS_XROOTD_CACHE" add_condor_vars_line ATLAS_XROOTD_CACHE "S" "-" "+" "Y" "Y" "+" So users can use env var without having to know where jobs actually land: # Job submitted from UCD lands at UCR: xrdcp root://uclhc-1.ucr.edu:4094//store/mc/runiifall15dr76/bulkgravtohhtohvvhbb_narrow_m-900_13tev- madgraph/aodsim/pu25nsdata2015v1_76x_mcrun2_asymptotic_v12-v1/10000/40b50f72-5bb4-e511- A31F-001517FB1B60.root. # Using env var instead xrdcp root://$cms_xrootd_cache//store/mc/runiifall15dr76/bulkgravtohhtohvvhbb_narrow_m-900_13tev- madgraph/aodsim/pu25nsdata2015v1_76x_mcrun2_asymptotic_v12-v1/10000/40b50f72-5bb4-e511- A31F-001517FB1B60.root. HTCondor Week 2016 13

Benefits of Condor-C Requires minimal change to existing batch config - We provide a drop-in file for /etc/condor/config.d Users continue using the submit host they are used to logging into - We just teach the magic lines to add to the submit file Makes user account creation on the brick unnecessary, improving security of DMZ since brick doubles as a data transfer node HTCondor Week 2016 14

Drop-in Condor-C config # track which site submitted from for accounting and gwms logic (matches GLIDIEN_Site) SUBMIT_SITE = UCSC" # submit attr defaults local = True site_local = True sdsc = False uc = False osg = False SUBMIT_SITE_ATTRS = local site_local sdsc uc osg SUBMIT_EXPRS = $(SUBMIT_EXPRS) SUBMIT_SITE $(SUBMIT_SITE_ATTRS) # ensure GSI is enabled for Condor-C SEC_CLIENT_AUTHENTICATION_METHODS = GSI, $(SEC_CLIENT_AUTHENTICATION_METHODS) # don t let GSI complain if submitting via brick local network IP GSI_SKIP_HOST_CHECK_CERT_REGEX = ^\/DC\=org\/DC\=opensciencegrid\/O\=Open\ Science\ Grid\/OU\=Services \/CN\=uclhc\-1\.ucsc\.edu$ HTCondor Week 2016 15

Traditional T3 Worker Worker Submit Host Worker Worker Local Batch HTCondor Week 2016 16

Uber Tier 3 OSG Comet Submit Host Local Batch Other UCs HTCondor Week 2016 17

UCSC Cache Stress Test Test jobs running xrdcp bring data in from CMS AAA data federation As expected initially we saturated the inbound at 10 gbps Over time network usage reduces as we access already fetched files from the disk cache Artificially increased slot count to 120 to push network limits Network choppiness and disk thrashing was caused by a subset of files xrootd was failing to fetch - O(1000) lingering TCP connections observed between cache and data servers long after clients disconnected from cache - We suspect a bug in the xrootd proxy cache when accessing broken files, more testing is needed HTCondor Week 2016 18

Attn HTCondor Devs! Surprises but not critical (Condor-C): - Unable to get Condor-C to use password auth Started thread with HTCondor support Not a big issue since our users already have grid certs, so we chose GSI - Condor-C doesn t use local network unless you give local hostname / ip in submit file - SUBMIT_EXPRS has to be set on submission side schedd variable is ignored on remote schedd side Improvements we definitely want: - Very interested in the generalization of monitoring metrics, we dropped Ganglia in favor of Influx DB Currently guilty of periodic condor-q s to parse running user job numbers (sorry Brian!) An it would be nice : - Any plans on implementing expression based flocking? (aka Edgar s wish list) HTCondor Week 2016 19

Conclusion LHC @ UC project utilizes the PRP network to enhance the T3s at each site by providing: - A unified way to submit locally and compute globally - The ability to decouple data placement from where the jobs run The central management of the services by dedicated admins at UCSD allows the local UC users to worry less about infrastructure maintenance and focus more on getting science done HTCondor Week 2016 20

One HTCondor pool to rule them all Questions? HTCondor Week 2016 21