Monitoring and Analytics With HTCondor Data

Similar documents
New Directions and BNL

Changing landscape of computing at BNL

@InfluxDB. David Norton 1 / 69

What s new in HTCondor? What s coming? HTCondor Week 2018 Madison, WI -- May 22, 2018

Cloud Computing. Summary

Monitoring Primer HTCondor Week 2017 Todd Tannenbaum Center for High Throughput Computing University of Wisconsin-Madison

The Art of Container Monitoring. Derek Chen

Time Series Live 2017

Developing Microsoft Azure Solutions (70-532) Syllabus

Monitor your containers with the Elastic Stack. Monica Sarbu

Developing Microsoft Azure Solutions (70-532) Syllabus

HTCondor on Titan. Wisconsin IceCube Particle Astrophysics Center. Vladimir Brik. HTCondor Week May 2018

CERN: LSF and HTCondor Batch Services

Monitoring system for geographically distributed datacenters based on Openstack. Gioacchino Vino

Shooting for the sky: Testing the limits of condor. HTCondor Week May 2015 Edgar Fajardo On behalf of OSG Software and Technology

Teraflops of Jupyter: A Notebook Based Analysis Portal at BNL

Developing Microsoft Azure Solutions (70-532) Syllabus

Visualize Your Data With Grafana Percona Live Daniel Lee - Software Engineer at Grafana Labs

Analytics Platform for ATLAS Computing Services

CS November 2017

Care and Feeding of HTCondor Cluster. Steven Timm European HTCondor Site Admins Meeting 8 December 2014

Performance Monitoring and Management of Microservices on Docker Ecosystem

UW-ATLAS Experiences with Condor

Monitoring Docker Containers with Splunk

Distributed CI: Scaling Jenkins on Mesos and Marathon. Roger Ignazio Puppet Labs, Inc. MesosCon 2015 Seattle, WA

PROOF-Condor integration for ATLAS

Linux Clusters Institute: Monitoring. Zhongtao Zhang, System Administrator, Holland Computing Center, University of Nebraska-Lincoln

CS November 2018

Data Analytics with HPC. Data Streaming

Search Engines and Time Series Databases

HTCondor overview. by Igor Sfiligoi, Jeff Dost (UCSD)

Monitoring MySQL with Prometheus & Grafana

One Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool. D. Mason for CMS Software & Computing

Monitor your infrastructure with the Elastic Beats. Monica Sarbu

Open Source Database Performance Optimization and Monitoring with PMM. Fernando Laudares, Vinicius Grippa, Michael Coburn Percona

HTCondor Week 2015: Implementing an HTCondor service at CERN

CHTC Policy and Configuration. Greg Thain HTCondor Week 2017

The SMACK Stack: Spark*, Mesos*, Akka, Cassandra*, Kafka* Elizabeth K. Dublin Apache Kafka Meetup, 30 August 2017.

Monitoring for IT Services and WLCG. Alberto AIMAR CERN-IT for the MONIT Team

Introducing the HTCondor-CE

The ATLAS EventIndex: Full chain deployment and first operation

HTCondor: Virtualization (without Virtual Machines)

Prototyping Data Intensive Apps: TrendingTopics.org

The InfluxDB-Grafana plugin for Fuel Documentation

Tutorial 4: Condor. John Watt, National e-science Centre

An update on the scalability limits of the Condor batch system

70-532: Developing Microsoft Azure Solutions

Elasticsearch & ATLAS Data Management. European Organization for Nuclear Research (CERN)

Building a Scalable Recommender System with Apache Spark, Apache Kafka and Elasticsearch

Open-Falcon A Distributed and High-Performance Monitoring System. Yao-Wei Ou & Lai Wei 2017/05/22

Bring Context To Your Machine Data With Hadoop, RDBMS & Splunk

Ingest. David Pilato, Developer Evangelist Paris, 31 Janvier 2017

glideinwms Training Glidein Internals How they work and why by Igor Sfiligoi, Jeff Dost (UCSD) glideinwms Training Glidein internals 1

Ingest. Aaron Mildenstein, Consulting Architect Tokyo Dec 14, 2017

Real-time monitoring Slurm jobs with InfluxDB September Carlos Fenoy García

Table 1 The Elastic Stack use cases Use case Industry or vertical market Operational log analytics: Gain real-time operational insight, reduce Mean Ti

Search and Time Series Databases

Qunar Performs Real-Time Data Analytics up to 300x Faster with Alluxio

What's new in Graphite 1.1. Denys FOSDEM 2018

MQ Monitoring on Cloud

Evolution of Cloud Computing in ATLAS

UK Tier-2 site evolution for ATLAS. Alastair Dewhurst

RIPE NCC Routing Information Service (RIS)

Effecient monitoring with Open source tools. Osman Ungur, github.com/o

The HTCondor CacheD. Derek Weitzel, Brian Bockelman University of Nebraska Lincoln

Look What I Can Do: Unorthodox Uses of HTCondor in the Open Science Grid

High-Energy Physics Data-Storage Challenges

Run your own Open source. (MMS) to avoid vendor lock-in. David Murphy MongoDB Practice Manager, Percona

What s new in HTCondor? What s coming? European HTCondor Workshop June 8, 2017

and the GridKa mass storage system Jos van Wezel / GridKa

LIME: A Framework for Lustre Global QoS Management. Li Xi DDN/Whamcloud Zeng Lingfang - JGU

70-532: Developing Microsoft Azure Solutions

MySQL Performance Improvements

Using Prometheus with InfluxDB for metrics storage

Topics. Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples

Application monitoring with BELK. Nishant Sahay, Sr. Architect Bhavani Ananth, Architect

GoDocker. A batch scheduling system with Docker containers

Firefox Crash Reporting.

Matchmaker Policies: Users and Groups HTCondor Week, Madison 2016

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

Trending with Purpose. Jason Dixon

Professional PostgreSQL monitoring made easy. Kaarel Moppel Kaarel Moppel

Monitoring HTCondor with the BigPanDA monitoring package

Talend Big Data Sandbox. Big Data Insights Cookbook

Road to Auto Scaling

5 Fundamental Strategies for Building a Data-centered Data Center

Flying HTCondor at 100gbps Over the Golden State

OSG Lessons Learned and Best Practices. Steven Timm, Fermilab OSG Consortium August 21, 2006 Site and Fabric Parallel Session

Monitoring MySQL Performance with Percona Monitoring and Management

Creating a Recommender System. An Elasticsearch & Apache Spark approach

Apache Storm. Hortonworks Inc Page 1

IBM Db2 Event Store Simplifying and Accelerating Storage and Analysis of Fast Data. IBM Db2 Event Store

Know Your Network. Computer Security Incident Response Center. Tracking Compliance Identifying and Mitigating Threats

DESY site report. HEPiX Spring 2016 at DESY. Yves Kemp, Peter van der Reest. Zeuthen,

Konstantin Shvachko, Hairong Kuang, Sanjay Radia, Robert Chansler Yahoo! Sunnyvale, California USA {Shv, Hairong, SRadia,

DC/OS Metrics. (formerly known as Project Ambrose) Application and Resource Metrics in DC/OS Enterprise. Nick Parker at..

Clouds in High Energy Physics

Graphite and Grafana

glideinwms UCSD Condor tunning by Igor Sfiligoi (UCSD) UCSD Jan 18th 2012 Condor Tunning 1

Distributed File Systems II

Transcription:

Monitoring and Analytics With HTCondor Data William Strecker-Kellogg RACF/SDCC @ BNL 1

RHIC/ATLAS Computing Facility (SDCC) Who are we? See our last two site reports from the HEPiX conference for a good overview of who we are and what we do. Recent news: running singularity for ATLAS jobs in production In short: support computing for RHIC collider at Brookhaven National Lab Condor pools for RHIC, 2 x ~16k core + smaller experiments In short: act as the Tier-1 for ATLAS in the US, storing data from CERN and distributing across US, as well as run large compute farm Host 90Pb of tape storage, dozens of Pb disk on Ceph, dcache, etc Runs around 18kCPU farm Increasing HPC presence, GPU-based Cluster + new OPA + KNL cluster 2

Batch Monitoring 3

Batch Monitoring Why? 1. Analytics 2. Accounting Is my farm occupied? Who exactly ran, when, where? With whom? Let's charge them per cpu-hour No jobs are starting! Send Alerts! Who's killing my storage!? Anathema to HTC Exactly how full were our resources? How cpu-efficient were we? 4

Batch Monitoring Why? Batch Systems, especially HTCondor, are great at producing vast streams of low-meaning data Wrangling it together to distill meaning / utility is a data-science challenge Condor cares who has what work and where it can be run. Really, really good at this. But. Plops bits and pieces of state history all over in log files on the schedd and startd side 5

State? Distributed. Schedds know jobs, startds know running jobs, periodically update schedd Collector knows little; just enough to query and match. Matching is stateless Only persistent pool-wide metadata (excluding offline ads) is Accounting data Kept in logfile where negotiator runs 6

Need for Accounting: Current Status Command digests history log, reads "***"-separated classad text Dumps to ever-growing mysql DB Carefully watches job status transitions for restarts and evictions, generates statistics about efficiency per-job and good(bad)put, etc... This is ugly. I did this. Years ago. Don't look at it. Doesn't this sound like a great case for logstash, Spark, etc... framework here...)? (insert trendy "big data" analytics 7

Accounting Analytics 8

Background of Monitoring Technologies 3 Components: Collect, Store, Display Display: Grafana Collect: Fifemon + own code Store: Graphite Additional aggregation / analysis layers Statsd Carbon-aggregator Elasticsearch 9

Graphite Architecture Four Components relay, cache, web & db DB: RRD-Like 1 file-per-metric (whisper) Define retentions based on metric regexp Directory structure is fast index Cluster Node: 1 x Relay -> N x Caches N x Caches -> 1 x DB 1 x Web <- DB + Caches Cache: holds points in memory on way to disk (carbon-cache) Configurable write-io profile, aggregates & groups metrics Inter-Node: Arbitrary architecture of relays (consistent hashing) E.G. 1 incoming relay -> N x cluster node head relay Web (graphite-web) WSGI App, own GUI + REST API Grafana talks REST Queries to one webapp fan out and merge across cluster 10

Graphite Architecture HAProxy can go here Arbitrary layers of relays + replication factors here... Grafana 11

Graphite Limitations Multiple dimensions in the data bloat the metric space, for example: "condor.<exp>.<slot>.{ram, cpu, etc...}" may have 10 metrics in 15k slots in 3 experiments "condor.<exp>.<user>.<schedd>.{ram, cpu...}" may have 10 metrics, 500 users across 10 schedds in 3 experiments 450k Metrics 150k Metrics Now take the cartesian product of these, and you'll soon run out of inodes on your graphite cluster 12

Graphite Limitations This sort of thing will drive us to look at multidimensional, tag-based TSDB InfluxDB recently went "open-core" and made clustering enterprise only ( ) So... OpenTSDB? Do we really want HDFS + HBASE just for condor monitoring? Prometheus? Ehh? Maybe? At this point, living with graphite sounds reasonable for us 13

Ancient (policy) History Mimicking "Queues" with HTCondor STAR and PHENIX experiments: different ways to classify jobs +Job_Type = "xyz" Owner = "special" +HighMem = true Flocking: HomeExperiment vs Experiment ATLAS AccountingGroups 14

Every site cares about monitoring a different aspect of their pool 15

HTCondor Monitoring Collects condor jobs from startds <experiment>.<slot-type>.<state>.<custom1 n>.<group>.<user> Schedd + Collector + Negotiator stats <experiment>.<daemon>.<stat-name> User Priorities + Usage <experiment>.<domain>.<group>.<user>.<usage prio > Tip: Keep wildcards above high-multiplicity points in the hierarchy 16

Custom Hierarchy Heavily modified Fifemon to support custom hierarchies of classads <experiment>.slots.<type>.<state>.<job-experiment>.<job-type>.<user>.num phenix.slots.static.claimed.phenix.crs.*.cpus phenix.slots.static.claimed.phenix.highmem.*.cpus 17

ard hbo Das tus Sta 18

PH E NI X Qu er y 19

Direct CGroup Monitoring Idea: mine the condor-generated cgroups for useful info to send to a TSDB Why? Isn't all the info in the condor_schedd? Schedd updates only periodically, incomplete info compared to cgroup Google's cadvisor? Go, embedded webserver, docker, ehhh My own solution? Small libcgroup-based script that sends data to graphite or statsd. 20

ATLAS CGroup Interface 21

CGroup Monitor Written in C, cmake RPM Reports to graphite via TCP or UDP Alert high memory / swap usage Alert low CPU-efficiency Can be extended to measure IO per device, etc Run under cron Also to StatsD Measurements per-job Note: do not look into the depths of libcgroup! puppet s fqdn_rand() for splay Integration: Collectd plugin? HTCondor? 22

CGroup Monitor Who is swapping? 23

CGroup Monitor How efficient is a job? 24

Conclusion Analytics is not accounting Graphite is a stable but effective choice of a TSDB Integration with HTCondor would be great Some extension done by me, happy to collaborate Custom cgroup monitoring proven useful for us Good performance, great query interface via Grafana Fifemon is a step in the right direction We're interested in the former for now Where to put it? Need elasticsearch or similar for ingesting startd / schedd history logs Area of current work for us 25

Thank You! Questions? Comments? We're hiring. Contact Eric Lancon or talk to myself or Tony Wong here this week. 26