BOSCO Architecture. Derek Weitzel University of Nebraska Lincoln
|
|
- Eustace Short
- 5 years ago
- Views:
Transcription
1 BOSCO Architecture Derek Weitzel University of Nebraska Lincoln
2 Goals We want an easy to use method for users to do computational research It should be easy to install, use, and maintain It should be simple for the user
3 Methods Use what s already at clusters - Their identity management - Their access methods Present a consistent interface to users If demand increases, expand organically, cluster to cluster
4 User Scenario 1 What they have: - A computer - Access to one cluster - Processing for their research What they want: - Simple job submission / management - Their processing to be completed now!
5 Technology Uses HTCondor job submission for user jobs Uses SSH to connect to clusters Uses Glite s BLAHP for interface to cluster scheduler Auto detection of remote cluster OS and appropriate BOSCO installation
6 Technology Bosco Submit Host NAT Bosco Submit Host SSH SSH PBS Cluster Login Node WN's
7 User Benefits from BOSCO 1. Throttled submission to remote cluster - Automatically detected 2. Job data transferred back to local computer after job completion 3. Do not care about remote OS version
8 User Scenario 2 What they have: - A computer - Access to one (or more) clusters - Processing for their research What they want: - Simple job submission / management - Their processing to be completed now!
9 Technology Everything as before plus Submit Glideins to remote clusters - Glideins are dynamic Condor worker nodes - Provides consistent interface for user jobs - Full output transferred back
10 User Benefits from BOSCO 2 1. Throttled submission to remote cluster - Automatically detected 2. Job data transferred back to local computer after job completion 3. Do not care about remote OS version 4. Transparent multi-cluster load balancing 5. Consistent interface to worker nodes 6. Ability to Flock remote Condor clusters
11 Bosco 2
12 Job Throttling In the current 1.1, we support detection and throttling of submitted jobs. Detects the number of jobs that can be submitted to a PBS cluster Uses Condor to throttle the number of jobs that can be submitted to that cluster
13 Job Data Transferred Job data is transferred back to submit host after job completion Important if further analysis is needed on the data
14 Mulit-Cluster Submitting glideins to multiple clusters at once Jobs can run at any cluster that starts the glideins Execution host matched after submission
15 Multi-OS Support Remote OS detected at install time BOSCO version installed from the cloud All OS s can communicate with each other through the GAHP protocol.
16 Requirements Requirements on clusters are limited - Running PBS, LSF, Condor, or SGE - Shared home file system - Outbound internet connectivity Requirements on submit host - For scenario 1, none - For scenario 2 Public IP address 1 port open (11000)
17 Compatibility Tested by Pegasus team to be compatible Can use Dagman workflow management If it can run on Condor, it can run on BOSCO
18 Bosco Breakdown What s it good for Personal Supercomputing assistant Submitting many jobs with same cores / memory Throttling submission to save the Scheduler Transferring data back and forth to YOUR workstation What it s not great at Submitting jobs with varying cores and memory requirements
19 Now Time for demo What could go wrong?
20 What just happened? Added 2 clusters - sandhills local EL6 condor cluster - ff-grid local EL5 PBS cluster Submitted jobs that ran at all 2 clusters, with no difference in the jobs.
21 What just happened? Factory submitted Glidein to cluster(s) PBS, SGE, Condor, LSF Cluster Worker Node BOSCO Submit Submit Glidein Login Node
22 What just happened? Cluster scheduled the Glidein on a worker node PBS, SGE, Condor, LSF Cluster Start Glidein Worker Node Glidein BOSCO Submit Login Node
23 What just happened? Glidein reports back to submit host for work PBS, SGE, Condor, LSF Cluster Ask For Job Worker Node Glidein BOSCO Submit Login Node
24 What just happened? BOSCO submit host starts a job on the cluster PBS, SGE, Condor, LSF Cluster Worker Node Glidein BOSCO Submit Login Node Job Start Job
25 BOSCO You can find out more at Now will take any questions
BOSCO Architecture. Derek Weitzel University of Nebraska Lincoln
BOSCO Architecture Derek Weitzel University of Nebraska Lincoln Goals We want an easy to use method for users to do computational research It should be easy to install, use, and maintain It should be simple
More informationBuilding Campus HTC Sharing Infrastructures. Derek Weitzel University of Nebraska Lincoln (Open Science Grid Hat)
Building Campus HTC Sharing Infrastructures Derek Weitzel University of Nebraska Lincoln (Open Science Grid Hat) HCC: Campus Grids Motivation We have 3 clusters in 2 cities. Our largest (4400 cores) is
More informationThe HTCondor CacheD. Derek Weitzel, Brian Bockelman University of Nebraska Lincoln
The HTCondor CacheD Derek Weitzel, Brian Bockelman University of Nebraska Lincoln Today s Talk Today s talk summarizes work for my a part of my PhD Dissertation Also, this work has been accepted to PDPTA
More informationPegasus Workflow Management System. Gideon Juve. USC Informa3on Sciences Ins3tute
Pegasus Workflow Management System Gideon Juve USC Informa3on Sciences Ins3tute Scientific Workflows Orchestrate complex, multi-stage scientific computations Often expressed as directed acyclic graphs
More informationIntroducing the HTCondor-CE
Introducing the HTCondor-CE CHEP 2015 Presented by Edgar Fajardo 1 Introduction In summer 2012, OSG performed an internal review of major software components, looking for strategic weaknesses. One highlighted
More informationEnabling Distributed Scientific Computing on the Campus
University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Computer Science and Engineering: Theses, Dissertations, and Student Research Computer Science and Engineering, Department
More informationglideinwms: Quick Facts
glideinwms: Quick Facts glideinwms is an open-source Fermilab Computing Sector product driven by CMS Heavy reliance on HTCondor from UW Madison and we work closely with them http://tinyurl.com/glideinwms
More informationVC3. Virtual Clusters for Community Computation. DOE NGNS PI Meeting September 27-28, 2017
VC3 Virtual Clusters for Community Computation DOE NGNS PI Meeting September 27-28, 2017 Douglas Thain, University of Notre Dame Rob Gardner, University of Chicago John Hover, Brookhaven National Lab A
More informationPegasus WMS Automated Data Management in Shared and Nonshared Environments
Pegasus WMS Automated Data Management in Shared and Nonshared Environments Mats Rynge USC Information Sciences Institute Pegasus Workflow Management System NSF funded project and developed
More informationCondor-G: HTCondor for grid submission. Jaime Frey (UW-Madison), Jeff Dost (UCSD)
Condor-G: HTCondor for grid submission Jaime Frey (UW-Madison), Jeff Dost (UCSD) Acknowledgement These slides are heavily based on the presentation Jaime Frey gave at UCSD in Feb 2011 http://www.t2.ucsd.edu/twiki2/bin/view/main/glideinfactory1111
More informationGetting Started with OSG Connect ~ an Interactive Tutorial ~
Getting Started with OSG Connect ~ an Interactive Tutorial ~ Emelie Harstad , Mats Rynge , Lincoln Bryant , Suchandra Thapa ,
More informationShooting for the sky: Testing the limits of condor. HTCondor Week May 2015 Edgar Fajardo On behalf of OSG Software and Technology
Shooting for the sky: Testing the limits of condor 21 May 2015 Edgar Fajardo On behalf of OSG Software and Technology 1 Acknowledgement Although I am the one presenting. This work is a product of a collaborative
More informationglideinwms architecture by Igor Sfiligoi, Jeff Dost (UCSD)
glideinwms architecture by Igor Sfiligoi, Jeff Dost (UCSD) Outline A high level overview of the glideinwms Description of the components 2 glideinwms from 10k feet 3 Refresher - HTCondor A Condor pool
More informationOne Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool. D. Mason for CMS Software & Computing
One Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool D. Mason for CMS Software & Computing 1 Going to try to give you a picture of the CMS HTCondor/ glideinwms global pool What s the use case
More informationGrid Compute Resources and Grid Job Management
Grid Compute Resources and Job Management March 24-25, 2007 Grid Job Management 1 Job and compute resource management! This module is about running jobs on remote compute resources March 24-25, 2007 Grid
More informationXSEDE High Throughput Computing Use Cases
XSEDE High Throughput Computing Use Cases 31 May 2013 Version 0.3 XSEDE HTC Use Cases Page 1 XSEDE HTC Use Cases Page 2 Table of Contents A. Document History B. Document Scope C. High Throughput Computing
More informationCorral: A Glide-in Based Service for Resource Provisioning
: A Glide-in Based Service for Resource Provisioning Gideon Juve USC Information Sciences Institute juve@usc.edu Outline Throughput Applications Grid Computing Multi-level scheduling and Glideins Example:
More informationHigh Throughput Urgent Computing
Condor Week 2008 High Throughput Urgent Computing Jason Cope jason.cope@colorado.edu Project Collaborators Argonne National Laboratory / University of Chicago Pete Beckman Suman Nadella Nick Trebon University
More informationGT 4.2.0: Community Scheduler Framework (CSF) System Administrator's Guide
GT 4.2.0: Community Scheduler Framework (CSF) System Administrator's Guide GT 4.2.0: Community Scheduler Framework (CSF) System Administrator's Guide Introduction This guide contains advanced configuration
More informationConnecting Restricted, High-Availability, or Low-Latency Resources to a Seamless Global Pool for CMS
Journal of Physics: Conference Series PAPER OPEN ACCESS Connecting Restricted, High-Availability, or Low-Latency Resources to a Seamless Global Pool for CMS To cite this article: J Balcas et al 2017 J.
More informationManaging large-scale workflows with Pegasus
Funded by the National Science Foundation under the OCI SDCI program, grant #0722019 Managing large-scale workflows with Pegasus Karan Vahi ( vahi@isi.edu) Collaborative Computing Group USC Information
More informationGrid Compute Resources and Job Management
Grid Compute Resources and Job Management How do we access the grid? Command line with tools that you'll use Specialised applications Ex: Write a program to process images that sends data to run on the
More information30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy
Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy Why the Grid? Science is becoming increasingly digital and needs to deal with increasing amounts of
More informationDistributed Caching Using the HTCondor CacheD
Int'l Conf Par and Dist Proc Tech and Appl PDPTA'15 341 Distributed Caching Using the Derek Weitzel, Brian Bockelman, and David Swanson Computer Science and Engineering University of Nebraska Lincoln Lincoln,
More informationCloud Computing. Up until now
Cloud Computing Lecture 4 and 5 Grid: 2012-2013 Introduction. Up until now Definition of Cloud Computing. Grid Computing: Schedulers: Condor SGE 1 Summary Core Grid: Toolkit Condor-G Grid: Conceptual Architecture
More informationglideinwms Training Glidein Internals How they work and why by Igor Sfiligoi, Jeff Dost (UCSD) glideinwms Training Glidein internals 1
Glidein Internals How they work and why by Igor Sfiligoi, Jeff Dost (UCSD) Glidein internals 1 Refresher glidein_startup the glidein_startup script configures and starts Condor on the worker node Glidein
More informationLi Yu. University of Notre Dame
Li Yu University of Notre Dame 1 Distributed systems are hard to use! An abstraction is a regular structure that can be efficiently scaled up to large problem sizes. We have implemented abstractions such
More informationThe Problem of Grid Scheduling
Grid Scheduling The Problem of Grid Scheduling Decentralised ownership No one controls the grid Heterogeneous composition Difficult to guarantee execution environments Dynamic availability of resources
More informationglite Grid Services Overview
The EPIKH Project (Exchange Programme to advance e-infrastructure Know-How) glite Grid Services Overview Antonio Calanducci INFN Catania Joint GISELA/EPIKH School for Grid Site Administrators Valparaiso,
More informationAdvanced School in High Performance and GRID Computing November Introduction to Grid computing.
1967-14 Advanced School in High Performance and GRID Computing 3-14 November 2008 Introduction to Grid computing. TAFFONI Giuliano Osservatorio Astronomico di Trieste/INAF Via G.B. Tiepolo 11 34131 Trieste
More informationOutline. ASP 2012 Grid School
Distributed Storage Rob Quick Indiana University Slides courtesy of Derek Weitzel University of Nebraska Lincoln Outline Storage Patterns in Grid Applications Storage
More informationA Virtual Comet. HTCondor Week 2017 May Edgar Fajardo On behalf of OSG Software and Technology
A Virtual Comet HTCondor Week 2017 May 3 2017 Edgar Fajardo On behalf of OSG Software and Technology 1 Working in Comet What my friends think I do What Instagram thinks I do What my boss thinks I do 2
More informationHPC and IT Issues Session Agenda. Deployment of Simulation (Trends and Issues Impacting IT) Mapping HPC to Performance (Scaling, Technology Advances)
HPC and IT Issues Session Agenda Deployment of Simulation (Trends and Issues Impacting IT) Discussion Mapping HPC to Performance (Scaling, Technology Advances) Discussion Optimizing IT for Remote Access
More informationWhat s new in HTCondor? What s coming? HTCondor Week 2018 Madison, WI -- May 22, 2018
What s new in HTCondor? What s coming? HTCondor Week 2018 Madison, WI -- May 22, 2018 Todd Tannenbaum Center for High Throughput Computing Department of Computer Sciences University of Wisconsin-Madison
More informationWork Queue + Python. A Framework For Scalable Scientific Ensemble Applications
Work Queue + Python A Framework For Scalable Scientific Ensemble Applications Peter Bui, Dinesh Rajan, Badi Abdul-Wahid, Jesus Izaguirre, Douglas Thain University of Notre Dame Distributed Computing Examples
More informationGrid Scheduling Architectures with Globus
Grid Scheduling Architectures with Workshop on Scheduling WS 07 Cetraro, Italy July 28, 2007 Ignacio Martin Llorente Distributed Systems Architecture Group Universidad Complutense de Madrid 1/38 Contents
More informationHTCondor with KRB/AFS Setup and first experiences on the DESY interactive batch farm
HTCondor with KRB/AFS Setup and first experiences on the DESY interactive batch farm Beyer Christoph & Finnern Thomas Madison (Wisconsin), May 2018 HTCondor week The Team and the Outline The Team Outline
More informationClouds: An Opportunity for Scientific Applications?
Clouds: An Opportunity for Scientific Applications? Ewa Deelman USC Information Sciences Institute Acknowledgements Yang-Suk Ki (former PostDoc, USC) Gurmeet Singh (former Ph.D. student, USC) Gideon Juve
More informationg-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.
g-eclipse A Framework for Accessing Grid Infrastructures Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.cy) EGEE Training the Trainers May 6 th, 2009 Outline Grid Reality The Problem g-eclipse
More informationDistributed Computing Framework. A. Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille
Distributed Computing Framework A. Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille EGI Webinar, 7 June 2016 Plan DIRAC Project Origins Agent based Workload Management System Accessible computing resources Data
More informationNew Directions and BNL
New Directions and HTCondor @ BNL USATLAS TIER-3 & NEW COMPUTING DIRECTIVES William Strecker-Kellogg RHIC/ATLAS Computing Facility (RACF) Brookhaven National Lab May 2016 RACF Overview 2 RHIC Collider
More informationVMs at a Tier-1 site. EGEE 09, Sander Klous, Nikhef
VMs at a Tier-1 site EGEE 09, 21-09-2009 Sander Klous, Nikhef Contents Introduction Who are we? Motivation Why are we interested in VMs? What are we going to do with VMs? Status How do we approach this
More informationChanging landscape of computing at BNL
Changing landscape of computing at BNL Shared Pool and New Users and Tools HTCondor Week May 2018 William Strecker-Kellogg Shared Pool Merging 6 HTCondor Pools into 1 2 What? Current Situation
More informationLogic Networks on the Grid: Handling 15 Million Jobs
Logic Networks on the Grid: Handling 15 Million Jobs Jan Bot, Delft Bioinformatics Lab 07-06-10 Delft University of Technology Challenge the future Overview Explanation of the application Challenges for
More informationHTCondor on Titan. Wisconsin IceCube Particle Astrophysics Center. Vladimir Brik. HTCondor Week May 2018
HTCondor on Titan Wisconsin IceCube Particle Astrophysics Center Vladimir Brik HTCondor Week May 2018 Overview of Titan Cray XK7 Supercomputer at Oak Ridge Leadership Computing Facility Ranked #5 by TOP500
More informationGanga The Job Submission Tool. WeiLong Ueng
Ganga The Job Submission Tool WeiLong Ueng wlueng@twgrid.org Objectives This tutorial gives users to understand Why require Ganga in Grid environment What advantages of Ganga The Architecture of Ganga
More informationAdaptive Co-Scheduler for Highly Dynamic Resources
University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Theses, Dissertations, & Student Research in Computer Electronics & Engineering Electrical & Computer Engineering, Department
More informationWLCG Lightweight Sites
WLCG Lightweight Sites Mayank Sharma (IT-DI-LCG) 3/7/18 Document reference 2 WLCG Sites Grid is a diverse environment (Various flavors of CE/Batch/WN/ +various preferred tools by admins for configuration/maintenance)
More informationGrid Programming: Concepts and Challenges. Michael Rokitka CSE510B 10/2007
Grid Programming: Concepts and Challenges Michael Rokitka SUNY@Buffalo CSE510B 10/2007 Issues Due to Heterogeneous Hardware level Environment Different architectures, chipsets, execution speeds Software
More informationWorkflow applications on EGI with WS-PGRADE. Peter Kacsuk and Zoltan Farkas MTA SZTAKI
applications on EGI with WS-PGRADE Peter Kacsuk and Zoltan Farkas MTA SZTAKI kacsuk@sztaki.hu 1 WS-PGRADE/gUSE Generic-purpose gateway framework Based on Liferay General purpose -oriented gateway framework
More informationSetup InstrucIons. If you need help with the setup, please put a red sicky note at the top of your laptop.
Setup InstrucIons Please complete these steps for the June 26 th workshop before the lessons start at 1:00 PM: h;p://hcc.unl.edu/june-workshop-setup#weekfour And make sure you can log into Crane. OS-specific
More informationThe glite middleware. Ariel Garcia KIT
The glite middleware Ariel Garcia KIT Overview Background The glite subsystems overview Security Information system Job management Data management Some (my) answers to your questions and random rumblings
More informationX Grid Engine. Where X stands for Oracle Univa Open Son of more to come...?!?
X Grid Engine Where X stands for Oracle Univa Open Son of more to come...?!? Carsten Preuss on behalf of Scientific Computing High Performance Computing Scheduler candidates LSF too expensive PBS / Torque
More informationFirst evaluation of the Globus GRAM Service. Massimo Sgaravatto INFN Padova
First evaluation of the Globus GRAM Service Massimo Sgaravatto INFN Padova massimo.sgaravatto@pd.infn.it Draft version release 1.0.5 20 June 2000 1 Introduction...... 3 2 Running jobs... 3 2.1 Usage examples.
More informationIvane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU
Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU Grid cluster at the Institute of High Energy Physics of TSU Authors: Arnold Shakhbatyan Prof. Zurab Modebadze Co-authors:
More informationUniversity of Alberta
University of Alberta PLACEHOLDER SCHEDULING FOR OVERLAY METACOMPUTING by Christopher James Pinchak A thesis submitted to the Faculty of Graduate Studies and Research in partial fulfillment of the requirements
More informationEvolution of the ATLAS PanDA Workload Management System for Exascale Computational Science
Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science T. Maeno, K. De, A. Klimentov, P. Nilsson, D. Oleynik, S. Panitkin, A. Petrosyan, J. Schovancova, A. Vaniachine,
More informationQueuing and Scheduling on Compute Clusters
Queuing and Scheduling on Compute Clusters Andrew Caird acaird@umich.edu Queuing and Scheduling on Compute Clusters p.1/17 The reason for me being here Give some queuing background Introduce some queuing
More informationChapter 2 Introduction to the WS-PGRADE/gUSE Science Gateway Framework
Chapter 2 Introduction to the WS-PGRADE/gUSE Science Gateway Framework Tibor Gottdank Abstract WS-PGRADE/gUSE is a gateway framework that offers a set of highlevel grid and cloud services by which interoperation
More informationPegasus. Pegasus Workflow Management System. Mats Rynge
Pegasus Pegasus Workflow Management System Mats Rynge rynge@isi.edu https://pegasus.isi.edu Automate Why workflows? Recover Automates complex, multi-stage processing pipelines Enables parallel, distributed
More informationGanga - a job management and optimisation tool. A. Maier CERN
Ganga - a job management and optimisation tool A. Maier CERN Overview What is Ganga Ganga Architecture Use Case: LHCb Use Case: Lattice QCD New features 2 Sponsors Ganga is an ATLAS/LHCb joint project
More informationConsiderations for a grid-based Physics Analysis Facility. Dietrich Liko
Considerations for a grid-based Physics Analysis Facility Dietrich Liko Introduction Aim of our grid activities is to enable physicists to do their work Latest GANGA developments PANDA Tier-3 Taskforce
More informationPODShell: Simplifying HPC in the Cloud Workflow
PODShell: Simplifying HPC in the Cloud Workflow June 2011 Penguin provides Linux HPC Solutions Linux Systems Servers Workstations Cluster Management Software HPC as a Service - Penguin on Demand Professional
More informationMONTE CARLO SIMULATION FOR RADIOTHERAPY IN A DISTRIBUTED COMPUTING ENVIRONMENT
The Monte Carlo Method: Versatility Unbounded in a Dynamic Computing World Chattanooga, Tennessee, April 17-21, 2005, on CD-ROM, American Nuclear Society, LaGrange Park, IL (2005) MONTE CARLO SIMULATION
More informationCloud Computing. Up until now
Cloud Computing Lectures 3 and 4 Grid Schedulers: Condor, Sun Grid Engine 2012-2013 Introduction. Up until now Definition of Cloud Computing. Grid Computing: Schedulers: Condor architecture. 1 Summary
More informationEFFICIENT SCHEDULING TECHNIQUES AND SYSTEMS FOR GRID COMPUTING
EFFICIENT SCHEDULING TECHNIQUES AND SYSTEMS FOR GRID COMPUTING By JANG-UK IN A DISSERTATION PRESENTED TO THE GRADUATE SCHOOL OF THE UNIVERSITY OF FLORIDA IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR
More informationThe University of Oxford campus grid, expansion and integrating new partners. Dr. David Wallom Technical Manager
The University of Oxford campus grid, expansion and integrating new partners Dr. David Wallom Technical Manager Outline Overview of OxGrid Self designed components Users Resources, adding new local or
More informationKestrel An XMPP-Based Framework for Many Task Computing Applications
Kestrel An XMPP-Based Framework for Many Task Computing Applications Lance Stout Mike Murphy Sebastien Goasguen HISTORY/PURPOSE Kestrel s Goals Lightweight / Easy to set up Run cross-platform without re-compiling
More informationHungarian Supercomputing Grid 1
Hungarian Supercomputing Grid 1 Péter Kacsuk MTA SZTAKI Victor Hugo u. 18-22, Budapest, HUNGARY www.lpds.sztaki.hu E-mail: kacsuk@sztaki.hu Abstract. The main objective of the paper is to describe the
More informationHTCondor overview. by Igor Sfiligoi, Jeff Dost (UCSD)
HTCondor overview by Igor Sfiligoi, Jeff Dost (UCSD) Acknowledgement These slides are heavily based on the presentation Todd Tannenbaum gave at CERN in Feb 2011 https://indico.cern.ch/event/124982/timetable/#20110214.detailed
More informationIoan Raicu. Everyone else. More information at: Background? What do you want to get out of this course?
Ioan Raicu More information at: http://www.cs.iit.edu/~iraicu/ Everyone else Background? What do you want to get out of this course? 2 Data Intensive Computing is critical to advancing modern science Applies
More informationWelcome to HTCondor Week #16. (year 31 of our project)
Welcome to HTCondor Week #16 (year 31 of our project) CHTC Team 2014 2 Driven by the potential of Distributed Computing to advance Scientific Discovery Claims for benefits provided by Distributed Processing
More informationXRAY Grid TO BE OR NOT TO BE?
XRAY Grid TO BE OR NOT TO BE? 1 I was not always a Grid sceptic! I started off as a grid enthusiast e.g. by insisting that Grid be part of the ESRF Upgrade Program outlined in the Purple Book : In this
More informationCluster Computing. Resource and Job Management for HPC 16/08/2010 SC-CAMP. ( SC-CAMP) Cluster Computing 16/08/ / 50
Cluster Computing Resource and Job Management for HPC SC-CAMP 16/08/2010 ( SC-CAMP) Cluster Computing 16/08/2010 1 / 50 Summary 1 Introduction Cluster Computing 2 About Resource and Job Management Systems
More informationWhat s new in HTCondor? What s coming? European HTCondor Workshop June 8, 2017
What s new in HTCondor? What s coming? European HTCondor Workshop June 8, 2017 Todd Tannenbaum Center for High Throughput Computing Department of Computer Sciences University of Wisconsin-Madison Release
More informationHOW TO SUBMIT AN ASSIGNMENT
HOW TO SUBMIT AN ASSIGNMENT Submitting an assignment on Blackboard is fast and can be performed in a few steps. The process outlined in this guide will ensure that a file can be submitted with minimal
More informationOn the Use of Cloud Computing for Scientific Workflows
On the Use of Cloud Computing for Scientific Workflows Christina Hoffa 1, Gaurang Mehta 2, Timothy Freeman 3, Ewa Deelman 2, Kate Keahey 3, Bruce Berriman 4, John Good 4 1 Indiana University, 2 University
More informationI Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011
I Tier-3 di CMS-Italia: stato e prospettive Claudio Grandi Workshop CCR GRID 2011 Outline INFN Perugia Tier-3 R&D Computing centre: activities, storage and batch system CMS services: bottlenecks and workarounds
More informationCERN: LSF and HTCondor Batch Services
Batch @ CERN: LSF and HTCondor Batch Services Iain Steers, Jérôme Belleman, Ulrich Schwickerath IT-PES-PS INFN Visit: Batch Batch @ CERN 2 Outline The Move Environment Grid Pilot Local Jobs Conclusion
More informationGrid Architectural Models
Grid Architectural Models Computational Grids - A computational Grid aggregates the processing power from a distributed collection of systems - This type of Grid is primarily composed of low powered computers
More informationALICE Grid Activities in US
ALICE Grid Activities in US 1 ALICE-USA Computing Project ALICE-USA Collaboration formed to focus on the ALICE EMCal project Construction, installation, testing and integration participating institutions
More informationSwiftrun i. Swiftrun
i Swiftrun ii Contents 1 Introduction 1 2 Running older Swift releases 1 2.1 sites.xml........................................................ 1 2.2 tc.data.........................................................
More informationFactory Ops Site Debugging
Factory Ops Site Debugging This talk shows detailed examples of how we debug site problems By Jeff Dost (UCSD) Factory Ops Site Debugging 1 Overview Validation Rundiff Held Waiting Pending Unmatched Factory
More informationFlying HTCondor at 100gbps Over the Golden State
Flying HTCondor at 100gbps Over the Golden State Jeff Dost (UCSD) HTCondor Week 2016 1 What is PRP? Pacific Research Platform: - 100 gbit network extending from Southern California to Washington - Interconnects
More informationGrid Mashups. Gluing grids together with Condor and BOINC
Grid Mashups Gluing grids together with Condor and BOINC, Artyom Sharov, Assaf Schuster, Dan Geiger Technion Israel Institute of Technology 1 Problem... 2 Problem... 3 Problem... 4 Parallelization From
More informationIntroduction to Grid Infrastructures
Introduction to Grid Infrastructures Stefano Cozzini 1 and Alessandro Costantini 2 1 CNR-INFM DEMOCRITOS National Simulation Center, Trieste, Italy 2 Department of Chemistry, Università di Perugia, Perugia,
More informationAutonomic Condor Clouds. David Wolinsky ACIS P2P Group University of Florida
Autonomic Condor Clouds David Wolinsky ACIS P2P Group University of Florida So What's the Big Deal Support connectivity across the Internet, in constrained locations, and with clouds Simplify packaging
More informationPegasus User Guide
Pegasus 4.8.0 User Guide Pegasus 4.8.0 User Guide Table of Contents 1. Introduction... 1 Overview and Features... 1 Workflow Gallery... 2 About this Document... 2 Document Formats (Web, PDF)... 2 2. Tutorial...
More informationBrief review of the HEPIX 2011 spring Darmstadt, 2-6 May
Brief review of the HEPIX 2011 spring Darmstadt, 2-6 May http://indico.cern.ch/conferencedisplay.py?confid=118192 Andrey Y Shevel 7 June 2011 Andrey Y Shevel 1 The presentation outlook HEPiX program Site
More informationSiemens PLM Software. HEEDS MDO Setting up a Windows-to- Linux Compute Resource.
Siemens PLM Software HEEDS MDO 2018.04 Setting up a Windows-to- Linux Compute Resource www.redcedartech.com. Contents Introduction 1 On Remote Machine B 2 Installing the SSH Server 2 Configuring the SSH
More informationBatch Services at CERN: Status and Future Evolution
Batch Services at CERN: Status and Future Evolution Helge Meinhard, CERN-IT Platform and Engineering Services Group Leader HTCondor Week 20 May 2015 20-May-2015 CERN batch status and evolution - Helge
More informationSimplified CICD with Jenkins and Git on the ZeroStack Platform
DATA SHEET Simplified CICD with Jenkins and Git on the ZeroStack Platform In the technical article we will walk through an end to end workflow of starting from virtually nothing and establishing a CICD
More informationAn Introduction to Virtualization and Cloud Technologies to Support Grid Computing
New Paradigms: Clouds, Virtualization and Co. EGEE08, Istanbul, September 25, 2008 An Introduction to Virtualization and Cloud Technologies to Support Grid Computing Distributed Systems Architecture Research
More informationCHTC Policy and Configuration. Greg Thain HTCondor Week 2017
CHTC Policy and Configuration Greg Thain HTCondor Week 2017 Image credit: flickr user shanelin cc Image credit: wikipedia CHTC Pool Mission CHTC Pool Mission To improve computational research on campus
More informationQuick Start Guide. Table of Contents
Quick Start Guide Table of Contents Account Registration... 2 Signup Request... 2 Account Activation... 4 Running FLOW-3D on POD... 9 Launching the GUI... 9 Running Simulations... 11 Collaborating with
More informationCREAM-WMS Integration
-WMS Integration Luigi Zangrando (zangrando@pd.infn.it) Moreno Marzolla (marzolla@pd.infn.it) JRA1 IT-CZ Cluster Meeting, Rome, 12 13/9/2005 1 Preliminaries We had a meeting in PD between Francesco G.,
More informationAutomatic Dependency Management for Scientific Applications on Clusters. Ben Tovar*, Nicholas Hazekamp, Nathaniel Kremer-Herman, Douglas Thain
Automatic Dependency Management for Scientific Applications on Clusters Ben Tovar*, Nicholas Hazekamp, Nathaniel Kremer-Herman, Douglas Thain Where users are Scientist says: "This demo task runs on my
More informationOpenSees on Teragrid
OpenSees on Teragrid Frank McKenna UC Berkeley OpenSees Parallel Workshop Berkeley, CA What isteragrid? An NSF sponsored computational science facility supported through a partnership of 13 institutions.
More informationAutoPyFactory: A Scalable Flexible Pilot Factory Implementation
ATL-SOFT-PROC-2012-045 22 May 2012 Not reviewed, for internal circulation only AutoPyFactory: A Scalable Flexible Pilot Factory Implementation J. Caballero 1, J. Hover 1, P. Love 2, G. A. Stewart 3 on
More informationCloudBATCH: A Batch Job Queuing System on Clouds with Hadoop and HBase. Chen Zhang Hans De Sterck University of Waterloo
CloudBATCH: A Batch Job Queuing System on Clouds with Hadoop and HBase Chen Zhang Hans De Sterck University of Waterloo Outline Introduction Motivation Related Work System Design Future Work Introduction
More information