Accelerating Science with High Throughput Computing (HTC)
|
|
- Pearl Williams
- 5 years ago
- Views:
Transcription
1 Accelerating Science with High Throughput Computing (HTC) Miron Livny Morgridge Institute Of Research Center for High Throughput Computing Computer Sciences Department University of Wisconsin-Madison
2 The process is made practical using Condor high-throughput computing (Thain et al. 2005), which enables O(10**2) update evaluations of the expensive simulations to be executed concurrently across workstations that would otherwise sit idle. This work was carried out at the University of Southampton condor pool, which is composed of approximately 1100 laboratory workstations and office PCs running various versions of the WINDOWS operating system.
3 GPro installation on Computational Nodes for HPC For High Performance Computing (HPC) it is required to install GPro_HPC_3.3.3.exe. Please read the updated installation instruction in the GPro User Manual. The recommended Condor version to use is v Condor downloads are available from the home page of the University of Wisconsin: downloads/
4 Genotype to Phenotype The current run will identify regions of the plant chromosomes that govern the growth and behavior of a seedling root that has been contrived to react to an external situation. The raw data from 1293 trials, one root per trial, images every 2 minutes for 8 hours. The total number of images is 311,613 (~265 GB), which were processed by Condor. The run will map the tip angle at each time points as a separate DAG. To assess the satistical significance, the data must be shuffled and rerun. Each of the DAGs runs 25,000 permutations of the genotype data. The run may take 60 days using ~ 14,000 CPU hours per day.
5
6 High Throughput Computing We first introduced the distinction between High Performance Computing (HPC) and High Throughput Computing (HTC) in a seminar at the NASA Goddard Flight Center in July of 1996 and a month later at the European Laboratory for Particle Physics (CERN). In June of 1997 HPCWire published an interview on High Throughput Computing.
7 Why HTC? For many experimental scientists, scientific progress and quality of research are strongly linked to computing throughput. In other words, they are less concerned about instantaneous computing power. Instead, what matters to them is the amount of computing they can harness over a month or a year --- they measure computing power in units of scenarios per day, wind patterns per week, instructions sets per month, or crystal configurations per year.
8 High Throughput Computing is a activity FLOPY (60*60*24*7*52)*FLOPS
9 Addison Snell, VP/GM, High Productivity Computing: Predictions for the New HPC Actionable Market Intelligence for High Productivity Computing October 2007
10 Use of Condor by the LIGO Scientific Collaboration Condor handles 10 s of millions of jobs per year running on the LDG, and up to 500k jobs per DAG. Condor standard universe check pointing widely used, saving us from having to manage this. At Caltech, 30 million jobs processed using 22.8 million CPU hrs. on 1324 CPUs in last 30 months. For example, to search 1 yr. of data for GWs from the inspiral of binary neutron star and black hole systems takes ~2 million jobs, and months to run on several thousand ~2.6 GHz nodes.
11 From: Stuart Anderson Date: February 28, :51:32 PM EST To: Condor-LIGO mailing list Subject: [CondorLIGO] Largest LIGO workflow Pete, Here are some numbers you ask about for LIGO's use of DAGs to manage large data analysis tasks broken down by the largest number of jobs managed in different categories: 1) DAG Instance--one condor_dagman process: 196,862. 2) DAG Workflow--launched from a single condor_submit_dag but may include multiple automatic sub- or spliced DAGs: 1,120,659. 3) DAG Analysis--multiple instances of condor_submit_dag to analyze a common dataset with results combined into a single coherent scientific result: 6,200,000. 4) DAG Total--sum over all instances of condor dagman run: O (100M). P.S. These are lower bounds as I did not perform an exhaustive survey/ search, but they are probably close. Thanks.
12 Grid Laboratory of Wisconsin (GLOW) HTC at the campus level Usage 04/04-04/10 114M Hours
13 Open Science Grid (OSG) HTC at the National Level
14 The Qiang Cui (QC) Group We develop and apply theoretical tools (electronic structure, nuclear dynamics and statistical mechanics) to biophysical problems (enzyme catalysis, bioenergy transduction and biomaterials etc.). Started to use the High Through Parallel Computing (HTPC) provided by CHTC in 05/09 and expended to OSG in 11/09
15 QC Group Comsumption
16 Impact of these cycles 1. Iron-Catalyzed Oxidation Intermediates Captured in A DNA Repair Monooxygenase, C. Yi, G. Jia, G. Hou, Q. Dai, G. Zheng, X. Jian, C. G. Yang, Q. Cui, and C. He, Science, Submitted 2. Disruption and formation of surface salt bridges are coupled to DNA binding in integration host factor (IHF): a computational analysis, L. Ma, M. T. Record, Jr., N. Sundlass, R. T. Raines and Q. Cui, J. Mol. Biol, Submitted 3. An implicit solvent model for SCC-DFTB with Charge-Dependent Radii, G. Hou, X. Zhu and Q. Cui, J. Chem. Theo. Comp., Submitted 4. Sequence-dependent interaction of $\beta$-peptides with membranes, J. Mondal, X. Zhu, Q. Cui and A. Yethiraj, J. Am. Chem. Soc., Submitted 5. A new coarse-grained model for water: The importance of electrostatic interactions, Z. Wu, Q. Cui and A. Yethiraj, J. Phys. Chem. Submitted 6. How does bone sialoprotein promote the nucleation of hydroxyapatite? A molecular dynamics study using model peptides of different conformations, Y. Yang, Q. Cui, and N. Sahai, {\it Langmuir}, Submitted 7. Preferential interactions between small solutes and the protein backbone: A computational analysis, L. Ma, L. Pegram, M. T. Record, Jr., Q. Cui, Biochem., 49, (2010) 8. Establishing effective simulation protocols for $\beta$- and $\alpha/\beta$-peptides. III. Molecular Mechanical (MM) model for a non-cyclic $\beta$-residue, X. Zhu, P. K\"onig, M. Hoffman, A. Yethiraj and Q. Cui, J. Comp. Chem., In press (DOI: /jcc.21493) 9. Curvature Generation and Pressure Profile in Membrane with lysolipids: Insights from coarse-grained simulations, J. Yoo and Q. Cui, Biophys. J. 97, (2009)
17 And what about managing energy?
18 From Forbes Magazine Open Source Energy Savings Dan Woods, Software for spreading work over huge collections of computers can be used to cut power costs. Condor supports all the operating systems a typical company or research institution would have and is rock solid in terms of stability and functions for its intended purpose, which is carving up work and sending it out to any number of computers for processing. 18
(HT)Condor - Past and Future
(HT)Condor - Past and Future Miron Livny John P. Morgridge Professor of Computer Science Wisconsin Institutes for Discovery University of Wisconsin-Madison חי has the value of 18 חי means alive Europe
More informationCondor Week 2010, Madison, WI April 2010
Condor Week 2010, Madison, WI April 2010 Todd Tannenbaum, Greg Thain, Zach Miller {tannenba, gthain, zmiller } @cs.wisc.edu Center for High Throughput Computing Department of Computer Sciences Agenda Brief
More informationLarge Scale in What Dimension?
Large Scale in What Dimension? Miron Livny Center for High Throughput Computing Computer Sciences Department University of Wisconsin-Madison It does (not) scale! What does it mean and should you care?
More informationDay 9: Introduction to CHTC
Day 9: Introduction to CHTC Suggested reading: Condor 7.7 Manual: http://www.cs.wisc.edu/condor/manual/v7.7/ Chapter 1: Overview Chapter 2: Users Manual (at most, 2.1 2.7) 1 Turn In Homework 2 Homework
More informationWelcome to HTCondor Week #16. (year 31 of our project)
Welcome to HTCondor Week #16 (year 31 of our project) CHTC Team 2014 2 Driven by the potential of Distributed Computing to advance Scientific Discovery Claims for benefits provided by Distributed Processing
More informationGrid Compute Resources and Job Management
Grid Compute Resources and Job Management How do we access the grid? Command line with tools that you'll use Specialised applications Ex: Write a program to process images that sends data to run on the
More informationRe-dock of Roscovitine Against Human Cyclin-Dependent Kinase 2 with Molegro Virtual Docker
Tutorial Re-dock of Roscovitine Against Human Cyclin-Dependent Kinase 2 with Molegro Virtual Docker Prof. Dr. Walter Filgueira de Azevedo Jr. walter@azevedolab.net azevedolab.net 1 Introduction In this
More informationClouds: An Opportunity for Scientific Applications?
Clouds: An Opportunity for Scientific Applications? Ewa Deelman USC Information Sciences Institute Acknowledgements Yang-Suk Ki (former PostDoc, USC) Gurmeet Singh (former Ph.D. student, USC) Gideon Juve
More informationGrid Computing a new tool for science
Grid Computing a new tool for science CERN, the European Organization for Nuclear Research Dr. Wolfgang von Rüden Wolfgang von Rüden, CERN, IT Department Grid Computing July 2006 CERN stands for over 50
More informationReliability Engineering Analysis of ATLAS Data Reprocessing Campaigns
Journal of Physics: Conference Series OPEN ACCESS Reliability Engineering Analysis of ATLAS Data Reprocessing Campaigns To cite this article: A Vaniachine et al 2014 J. Phys.: Conf. Ser. 513 032101 View
More informationBig Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback
Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy Texas A&M Big Data Workshop October 2011 January 2015, Texas A&M University Research Topics Seminar 1 Outline Overview of
More informationUW-ATLAS Experiences with Condor
UW-ATLAS Experiences with Condor M.Chen, A. Leung, B.Mellado Sau Lan Wu and N.Xu Paradyn / Condor Week, Madison, 05/01/08 Outline Our first success story with Condor - ATLAS production in 2004~2005. CRONUS
More informationA Fully Automated Faulttolerant. Distributed Video Processing and Off site Replication
A Fully Automated Faulttolerant System for Distributed Video Processing and Off site Replication George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison June 2004 What is the talk about?
More informationConference The Data Challenges of the LHC. Reda Tafirout, TRIUMF
Conference 2017 The Data Challenges of the LHC Reda Tafirout, TRIUMF Outline LHC Science goals, tools and data Worldwide LHC Computing Grid Collaboration & Scale Key challenges Networking ATLAS experiment
More informationPegasus WMS Automated Data Management in Shared and Nonshared Environments
Pegasus WMS Automated Data Management in Shared and Nonshared Environments Mats Rynge USC Information Sciences Institute Pegasus Workflow Management System NSF funded project and developed
More informationDay 11: Workflows with DAGMan
Day 11: Workflows with DAGMan Suggested reading: Condor 7.7 Manual: http://www.cs.wisc.edu/condor/manual/v7.7/ Section 2.10: DAGMan Applications Chapter 9: condor_submit_dag 1 Turn In Homework 2 Homework
More informationProfiling Grid Data Transfer Protocols and Servers. George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA
Profiling Grid Data Transfer Protocols and Servers George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA Motivation Scientific experiments are generating large amounts of data Education
More informationCSCS CERN videoconference CFD applications
CSCS CERN videoconference CFD applications TS/CV/Detector Cooling - CFD Team CERN June 13 th 2006 Michele Battistin June 2006 CERN & CFD Presentation 1 TOPICS - Some feedback about already existing collaboration
More informationEvolution of the ATLAS PanDA Workload Management System for Exascale Computational Science
Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science T. Maeno, K. De, A. Klimentov, P. Nilsson, D. Oleynik, S. Panitkin, A. Petrosyan, J. Schovancova, A. Vaniachine,
More informationScientific Workflows and Cloud Computing. Gideon Juve USC Information Sciences Institute
Scientific Workflows and Cloud Computing Gideon Juve USC Information Sciences Institute gideon@isi.edu Scientific Workflows Loosely-coupled parallel applications Expressed as directed acyclic graphs (DAGs)
More informationNAMD Serial and Parallel Performance
NAMD Serial and Parallel Performance Jim Phillips Theoretical Biophysics Group Serial performance basics Main factors affecting serial performance: Molecular system size and composition. Cutoff distance
More informationManaging large-scale workflows with Pegasus
Funded by the National Science Foundation under the OCI SDCI program, grant #0722019 Managing large-scale workflows with Pegasus Karan Vahi ( vahi@isi.edu) Collaborative Computing Group USC Information
More informationPegasus. Pegasus Workflow Management System. Mats Rynge
Pegasus Pegasus Workflow Management System Mats Rynge rynge@isi.edu https://pegasus.isi.edu Automate Why workflows? Recover Automates complex, multi-stage processing pipelines Enables parallel, distributed
More informationCondor and BOINC. Distributed and Volunteer Computing. Presented by Adam Bazinet
Condor and BOINC Distributed and Volunteer Computing Presented by Adam Bazinet Condor Developed at the University of Wisconsin-Madison Condor is aimed at High Throughput Computing (HTC) on collections
More informationTypically applied in clusters and grids Loosely-coupled applications with sequential jobs Large amounts of computing for long periods of times
Typically applied in clusters and grids Loosely-coupled applications with sequential jobs Large amounts of computing for long periods of times Measured in operations per month or years 2 Bridge the gap
More informationInsight: that s for NSA Decision making: that s for Google, Facebook. so they find the best way to push out adds and products
What is big data? Big data is high-volume, high-velocity and high-variety information assets that demand cost-effective, innovative forms of information processing for enhanced insight and decision making.
More informationIntroduction to Grid Computing
Milestone 2 Include the names of the papers You only have a page be selective about what you include Be specific; summarize the authors contributions, not just what the paper is about. You might be able
More informationUsing Alluxio to Improve the Performance and Consistency of HDFS Clusters
ARTICLE Using Alluxio to Improve the Performance and Consistency of HDFS Clusters Calvin Jia Software Engineer at Alluxio Learn how Alluxio is used in clusters with co-located compute and storage to improve
More informationOn-demand provisioning of HEP compute resources on cloud sites and shared HPC centers
On-demand provisioning of HEP compute resources on cloud sites and shared HPC centers CHEP 2016 - San Francisco, United States of America Gunther Erli, Frank Fischer, Georg Fleig, Manuel Giffels, Thomas
More informationMissile External Aerodynamics Using Star-CCM+ Star European Conference 03/22-23/2011
Missile External Aerodynamics Using Star-CCM+ Star European Conference 03/22-23/2011 StarCCM_StarEurope_2011 4/6/11 1 Overview 2 Role of CFD in Aerodynamic Analyses Classical aerodynamics / Semi-Empirical
More information3D Printing: Chemistry and Crystal Structures
3D Printing: Chemistry and Crystal Structures Ian Bruno - Director, Strategic Partnerships The Cambridge Crystallographic Data Centre @ijbruno @ccdc_cambridge Science of 3D Printing SLA 2015 Annual Conference,
More informationScientific data processing at global scale The LHC Computing Grid. fabio hernandez
Scientific data processing at global scale The LHC Computing Grid Chengdu (China), July 5th 2011 Who I am 2 Computing science background Working in the field of computing for high-energy physics since
More informationSynonymous with supercomputing Tightly-coupled applications Implemented using Message Passing Interface (MPI) Large of amounts of computing for short
Synonymous with supercomputing Tightly-coupled applications Implemented using Message Passing Interface (MPI) Large of amounts of computing for short periods of time Usually requires low latency interconnects
More informationAn update on the scalability limits of the Condor batch system
An update on the scalability limits of the Condor batch system D Bradley 1, T St Clair 1, M Farrellee 1, Z Guo 1, M Livny 1, I Sfiligoi 2, T Tannenbaum 1 1 University of Wisconsin, Madison, WI, USA 2 University
More informationDetached Eddy Simulation Analysis of a Transonic Rocket Booster for Steady & Unsteady Buffet Loads
Detached Eddy Simulation Analysis of a Transonic Rocket Booster for Steady & Unsteady Buffet Loads Matt Knapp Chief Aerodynamicist TLG Aerospace, LLC Presentation Overview Introduction to TLG Aerospace
More informationMonte Carlo modelling and applications to imaging
Monte Carlo modelling and applications to imaging The Monte Carlo method is a method to obtain a result through repeated random sampling of the outcome of a system. One of the earliest applications, in
More informationMOLECULAR VISUALIZATION LAB USING PYMOL a supplement to Chapter 11. Please complete this tutorial before coming to your lab section
MOLECULAR VISUALIZATION LAB USING PYMOL a supplement to Chapter 11 Please complete this tutorial before coming to your lab section (Adapted from Dr. Vardar-Ulu Fall 2015) Before coming to your lab section
More informationReduction of Workflow Resource Consumption Using a Density-based Clustering Model
Reduction of Workflow Resource Consumption Using a Density-based Clustering Model Qimin Zhang, Nathaniel Kremer-Herman, Benjamin Tovar, and Douglas Thain WORKS Workshop at Supercomputing 2018 The Cooperative
More informationHigh Throughput Computing and Sampling Issues for Optimization in Radiotherapy
High Throughput Computing and Sampling Issues for Optimization in Radiotherapy Michael C. Ferris, University of Wisconsin Alexander Meeraus, GAMS Development Corporation Optimization Days, Montreal, May
More informationCustomer Success Story Los Alamos National Laboratory
Customer Success Story Los Alamos National Laboratory Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory Case Study June 2010 Highlights First Petaflop
More informationSoftware composition for scientific workflows
Software composition for scientific workflows Johan Montagnat CNRS / UNS, I3S, MODALIS April 17, 2009 http://gwendia.polytech.unice.fr Grid infrastructures Scientific production Europe: EGEE (www.eu-egee.org),
More information) I R L Press Limited, Oxford, England. The protein identification resource (PIR)
Volume 14 Number 1 Volume 1986 Nucleic Acids Research 14 Number 1986 Nucleic Acids Research The protein identification resource (PIR) David G.George, Winona C.Barker and Lois T.Hunt National Biomedical
More informationAgenda. Sill, Alan Texas Tech University Workshop Introduction 8:30 AM Recently Released OGF Standards; US Federal and Other Roadmap Efforts
Workshop on High Performance Applications of Cloud and Grid Tools Agenda Sill, Alan Texas Tech University Workshop Introduction 8:30 AM Sill, Alan Texas Tech University Recently Released OGF Standards;
More informationHDF Update. Elena Pourmal The HDF Group. October 16, 2008 IDL User Group Meeting 1
HDF Update Elena Pourmal The HDF Group October 16, 2008 IDL User Group Meeting 1 The HDF Group The HDF Group is a not-for-profit company with its mission focused on the support and growth of the HDF technologies
More informationWhat s new in HTCondor? What s coming? HTCondor Week 2018 Madison, WI -- May 22, 2018
What s new in HTCondor? What s coming? HTCondor Week 2018 Madison, WI -- May 22, 2018 Todd Tannenbaum Center for High Throughput Computing Department of Computer Sciences University of Wisconsin-Madison
More informationCarelyn Campbell, Ben Blaiszik, Laura Bartolo. November 1, 2016
Carelyn Campbell, Ben Blaiszik, Laura Bartolo November 1, 2016 Data Landscape Collaboration Tools (e.g. Google Drive, DropBox, Sharepoint, Github, MatIN) Data Sharing Communities (e.g. Dryad, FigShare,
More informationIoan Raicu Distributed Systems Laboratory Computer Science Department University of Chicago
Running 1 Million Jobs in 10 Minutes via the Falkon Fast and Light-weight Ioan Raicu Distributed Systems Laboratory Computer Science Department University of Chicago In Collaboration with: Ian Foster,
More informationPART I - Fundamentals of Parallel Computing
PART I - Fundamentals of Parallel Computing Objectives What is scientific computing? The need for more computing power The need for parallel computing and parallel programs 1 What is scientific computing?
More informationCondor and Workflows: An Introduction. Condor Week 2011
Condor and Workflows: An Introduction Condor Week 2011 Kent Wenger Condor Project Computer Sciences Department University of Wisconsin-Madison Outline > Introduction/motivation > Basic DAG concepts > Running
More informationPart IV. Workflow Mapping and Execution in Pegasus. (Thanks to Ewa Deelman)
AAAI-08 Tutorial on Computational Workflows for Large-Scale Artificial Intelligence Research Part IV Workflow Mapping and Execution in Pegasus (Thanks to Ewa Deelman) 1 Pegasus-Workflow Management System
More informationImproving Network Infrastructure to Enable Large Scale Scientific Data Flows and Collaboration (Award # ) Klara Jelinkova Joseph Ghobrial
Improving Network Infrastructure to Enable Large Scale Scientific Data Flows and Collaboration (Award # 1659348) Klara Jelinkova Joseph Ghobrial NSF Campus Cyberinfrastructure PI and Cybersecurity Innovation
More informationOverview and Introduction to Scientific Visualization. Texas Advanced Computing Center The University of Texas at Austin
Overview and Introduction to Scientific Visualization Texas Advanced Computing Center The University of Texas at Austin Scientific Visualization The purpose of computing is insight not numbers. -- R. W.
More informationTESLA P100 PERFORMANCE GUIDE. HPC and Deep Learning Applications
TESLA P PERFORMANCE GUIDE HPC and Deep Learning Applications MAY 217 TESLA P PERFORMANCE GUIDE Modern high performance computing (HPC) data centers are key to solving some of the world s most important
More informationSolving Hard Integer Programs with MW
Solving Hard Integer Programs with MW Jeff Linderoth ISE Department COR@L Lab Lehigh University jtl3@lehigh.edu 2007 Condor Jamboree Madison, WI May 2, 2007 Thanks! NSF OCI-0330607, CMMI-0522796, DOE DE-FG02-05ER25694
More informationOne Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool. D. Mason for CMS Software & Computing
One Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool D. Mason for CMS Software & Computing 1 Going to try to give you a picture of the CMS HTCondor/ glideinwms global pool What s the use case
More informationA Cloud-based Dynamic Workflow for Mass Spectrometry Data Analysis
A Cloud-based Dynamic Workflow for Mass Spectrometry Data Analysis Ashish Nagavaram, Gagan Agrawal, Michael A. Freitas, Kelly H. Telu The Ohio State University Gaurang Mehta, Rajiv. G. Mayani, Ewa Deelman
More informationGrid Compute Resources and Grid Job Management
Grid Compute Resources and Job Management March 24-25, 2007 Grid Job Management 1 Job and compute resource management! This module is about running jobs on remote compute resources March 24-25, 2007 Grid
More informationA Virtual Comet. HTCondor Week 2017 May Edgar Fajardo On behalf of OSG Software and Technology
A Virtual Comet HTCondor Week 2017 May 3 2017 Edgar Fajardo On behalf of OSG Software and Technology 1 Working in Comet What my friends think I do What Instagram thinks I do What my boss thinks I do 2
More informationData services for LHC computing
Data services for LHC computing SLAC 1 Xavier Espinal on behalf of IT/ST DAQ to CC 8GB/s+4xReco Hot files Reliable Fast Processing DAQ Feedback loop WAN aware Tier-1/2 replica, multi-site High throughout
More informationMaximize automotive simulation productivity with ANSYS HPC and NVIDIA GPUs
Presented at the 2014 ANSYS Regional Conference- Detroit, June 5, 2014 Maximize automotive simulation productivity with ANSYS HPC and NVIDIA GPUs Bhushan Desam, Ph.D. NVIDIA Corporation 1 NVIDIA Enterprise
More informationMaking Condor Safer with A Collaborative Marketplace for Continuous Software Assurance Brooklin Gore, Chief Operations Officer info@cosalab.org http://swamp.cosalab.org U.S. Department of Homeland Security
More informationResearch on Evaluation Method of Product Style Semantics Based on Neural Network
Research Journal of Applied Sciences, Engineering and Technology 6(23): 4330-4335, 2013 ISSN: 2040-7459; e-issn: 2040-7467 Maxwell Scientific Organization, 2013 Submitted: September 28, 2012 Accepted:
More informationRecent Advances in Modelling Wind Parks in STAR CCM+ Steve Evans
Recent Advances in Modelling Wind Parks in STAR CCM+ Steve Evans Introduction Company STAR-CCM+ Agenda Wind engineering at CD-adapco STAR-CCM+ & EnviroWizard Developments for Offshore Simulation CD-adapco:
More informationTESLA P100 PERFORMANCE GUIDE. Deep Learning and HPC Applications
TESLA P PERFORMANCE GUIDE Deep Learning and HPC Applications SEPTEMBER 217 TESLA P PERFORMANCE GUIDE Modern high performance computing (HPC) data centers are key to solving some of the world s most important
More informationTechnical Brief: Specifying a PC for Mascot
Technical Brief: Specifying a PC for Mascot Matrix Science 8 Wyndham Place London W1H 1PP United Kingdom Tel: +44 (0)20 7723 2142 Fax: +44 (0)20 7725 9360 info@matrixscience.com http://www.matrixscience.com
More informationOverview of ATLAS PanDA Workload Management
Overview of ATLAS PanDA Workload Management T. Maeno 1, K. De 2, T. Wenaus 1, P. Nilsson 2, G. A. Stewart 3, R. Walker 4, A. Stradling 2, J. Caballero 1, M. Potekhin 1, D. Smith 5, for The ATLAS Collaboration
More informationImprovements to the SHDOM Radiative Transfer Modeling Package
Improvements to the SHDOM Radiative Transfer Modeling Package K. F. Evans University of Colorado Boulder, Colorado W. J. Wiscombe National Aeronautics and Space Administration Goddard Space Flight Center
More informationProject Blackbird. U"lizing Condor and HTC to address archiving online courses at Clemson on a weekly basis. Sam Hoover
Project Blackbird U"lizing Condor and HTC to address archiving online courses at Clemson on a weekly basis Sam Hoover shoover@clemson.edu 1 Project Blackbird Blackboard at Clemson End of Semester archives
More informationOzenCloud Case Studies
OzenCloud Case Studies Case Studies, April 20, 2015 ANSYS in the Cloud Case Studies: Aerodynamics & fluttering study on an aircraft wing using fluid structure interaction 1 Powered by UberCloud http://www.theubercloud.com
More informationSTORK: Making Data Placement a First Class Citizen in the Grid
STORK: Making Data Placement a First Class Citizen in the Grid Tevfik Kosar University of Wisconsin-Madison May 25 th, 2004 CERN Need to move data around.. TB PB TB PB While doing this.. Locate the data
More informationGrid Mashups. Gluing grids together with Condor and BOINC
Grid Mashups Gluing grids together with Condor and BOINC, Artyom Sharov, Assaf Schuster, Dan Geiger Technion Israel Institute of Technology 1 Problem... 2 Problem... 3 Problem... 4 Parallelization From
More informationBig Data - Some Words BIG DATA 8/31/2017. Introduction
BIG DATA Introduction Big Data - Some Words Connectivity Social Medias Share information Interactivity People Business Data Data mining Text mining Business Intelligence 1 What is Big Data Big Data means
More informationPegasus. Automate, recover, and debug scientific computations. Mats Rynge https://pegasus.isi.edu
Pegasus Automate, recover, and debug scientific computations. Mats Rynge rynge@isi.edu https://pegasus.isi.edu Why Pegasus? Automates complex, multi-stage processing pipelines Automate Enables parallel,
More informationHigh Throughput Urgent Computing
Condor Week 2008 High Throughput Urgent Computing Jason Cope jason.cope@colorado.edu Project Collaborators Argonne National Laboratory / University of Chicago Pete Beckman Suman Nadella Nick Trebon University
More informationXSEDE High Throughput Computing Use Cases
XSEDE High Throughput Computing Use Cases 31 May 2013 Version 0.3 XSEDE HTC Use Cases Page 1 XSEDE HTC Use Cases Page 2 Table of Contents A. Document History B. Document Scope C. High Throughput Computing
More informationStorage and Compute Resource Management via DYRE, 3DcacheGrid, and CompuStore
Storage and Compute Resource Management via DYRE, 3DcacheGrid, and CompuStore Ioan Raicu Distributed Systems Laboratory Computer Science Department University of Chicago DSL Seminar November st, 006 Analysis
More informationIntroduction & Motivation Problem Statement Proposed Work Evaluation Conclusions Future Work
Introduction & Motivation Problem Statement Proposed Work Evaluation Conclusions Future Work Introduction & Motivation Problem Statement Proposed Work Evaluation Conclusions Future Work Today (2014):
More informationShaking-and-Baking on a Grid
Shaking-and-Baking on a Grid Russ Miller & Mark Green Center for Computational Research, SUNY-Buffalo Hauptman-Woodward Medical Inst NSF ITR ACI-02-04918 University at Buffalo The State University of New
More informationBuilding Campus HTC Sharing Infrastructures. Derek Weitzel University of Nebraska Lincoln (Open Science Grid Hat)
Building Campus HTC Sharing Infrastructures Derek Weitzel University of Nebraska Lincoln (Open Science Grid Hat) HCC: Campus Grids Motivation We have 3 clusters in 2 cities. Our largest (4400 cores) is
More informationPegasus. Automate, recover, and debug scientific computations. Mats Rynge
Pegasus Automate, recover, and debug scientific computations. Mats Rynge rynge@isi.edu https://pegasus.isi.edu Why Pegasus? Automates complex, multi-stage processing pipelines Automate Enables parallel,
More informationUsing GPUs to compute the multilevel summation of electrostatic forces
Using GPUs to compute the multilevel summation of electrostatic forces David J. Hardy Theoretical and Computational Biophysics Group Beckman Institute for Advanced Science and Technology University of
More informationWhat are Clusters? Why Clusters? - a Short History
What are Clusters? Our definition : A parallel machine built of commodity components and running commodity software Cluster consists of nodes with one or more processors (CPUs), memory that is shared by
More informationHigh Throughput WAN Data Transfer with Hadoop-based Storage
High Throughput WAN Data Transfer with Hadoop-based Storage A Amin 2, B Bockelman 4, J Letts 1, T Levshina 3, T Martin 1, H Pi 1, I Sfiligoi 1, M Thomas 2, F Wuerthwein 1 1 University of California, San
More informationHigh Performance Computing Course Notes Course Administration
High Performance Computing Course Notes 2009-2010 2010 Course Administration Contacts details Dr. Ligang He Home page: http://www.dcs.warwick.ac.uk/~liganghe Email: liganghe@dcs.warwick.ac.uk Office hours:
More informationA Study of High Performance Computing and the Cray SV1 Supercomputer. Michael Sullivan TJHSST Class of 2004
A Study of High Performance Computing and the Cray SV1 Supercomputer Michael Sullivan TJHSST Class of 2004 June 2004 0.1 Introduction A supercomputer is a device for turning compute-bound problems into
More informationBuilding a Virtualized Desktop Grid. Eric Sedore
Building a Virtualized Desktop Grid Eric Sedore essedore@syr.edu Why create a desktop grid? One prong of an three pronged strategy to enhance research infrastructure on campus (physical hosting, HTC grid,
More informationSriram Krishnan
A Web Services Based Architecture for Biomedical Applications Sriram Krishnan sriram@sdsc.edu Goals Enabling integration across multi-scale biomedical applications Leveraging geographically distributed,
More informationIntroduction to National Supercomputing Centre in Guangzhou and Opportunities for International Collaboration
Exascale Applications and Software Conference 21st 23rd April 2015, Edinburgh, UK Introduction to National Supercomputing Centre in Guangzhou and Opportunities for International Collaboration Xue-Feng
More informationBuilding the International Data Placement Lab. Greg Thain
Building the International Data Placement Lab Greg Thain What is the IDPL? Overview How we built it Examples of use 2 Who is the IDPL? Phil Papadapolus -- UCSD Miron Livny -- Wisconsin Collaborators in
More informationData oriented job submission scheme for the PHENIX user analysis in CCJ
Journal of Physics: Conference Series Data oriented job submission scheme for the PHENIX user analysis in CCJ To cite this article: T Nakamura et al 2011 J. Phys.: Conf. Ser. 331 072025 Related content
More informationCernVM-FS beyond LHC computing
CernVM-FS beyond LHC computing C Condurache, I Collier STFC Rutherford Appleton Laboratory, Harwell Oxford, Didcot, OX11 0QX, UK E-mail: catalin.condurache@stfc.ac.uk Abstract. In the last three years
More informationIEPSAS-Kosice: experiences in running LCG site
IEPSAS-Kosice: experiences in running LCG site Marian Babik 1, Dusan Bruncko 2, Tomas Daranyi 1, Ladislav Hluchy 1 and Pavol Strizenec 2 1 Department of Parallel and Distributed Computing, Institute of
More informationThe JINR Tier1 Site Simulation for Research and Development Purposes
EPJ Web of Conferences 108, 02033 (2016) DOI: 10.1051/ epjconf/ 201610802033 C Owned by the authors, published by EDP Sciences, 2016 The JINR Tier1 Site Simulation for Research and Development Purposes
More informationHigh Performance Computing on MapReduce Programming Framework
International Journal of Private Cloud Computing Environment and Management Vol. 2, No. 1, (2015), pp. 27-32 http://dx.doi.org/10.21742/ijpccem.2015.2.1.04 High Performance Computing on MapReduce Programming
More informationAccelerating the Scientific Exploration Process with Kepler Scientific Workflow System
Accelerating the Scientific Exploration Process with Kepler Scientific Workflow System Jianwu Wang, Ilkay Altintas Scientific Workflow Automation Technologies Lab SDSC, UCSD project.org UCGrid Summit,
More informationDAGMan workflow. Kumaran Baskaran
DAGMan workflow Kumaran Baskaran NMRbox summer workshop June 26-29,2017 what is a workflow? Laboratory Publication Workflow management Softwares Workflow management Data Softwares Workflow management
More informationMaking Supercomputing More Available and Accessible Windows HPC Server 2008 R2 Beta 2 Microsoft High Performance Computing April, 2010
Making Supercomputing More Available and Accessible Windows HPC Server 2008 R2 Beta 2 Microsoft High Performance Computing April, 2010 Windows HPC Server 2008 R2 Windows HPC Server 2008 R2 makes supercomputing
More informationOptimizing Parallel Access to the BaBar Database System Using CORBA Servers
SLAC-PUB-9176 September 2001 Optimizing Parallel Access to the BaBar Database System Using CORBA Servers Jacek Becla 1, Igor Gaponenko 2 1 Stanford Linear Accelerator Center Stanford University, Stanford,
More informationThe Bulk Synchronous Parallel Model (PSC 1.2) Lecture 1.2 Bulk Synchronous Parallel Model p.1
The Bulk Synchronous Parallel Model (PSC 1.2) Lecture 1.2 Bulk Synchronous Parallel Model p.1 What is a parallel computer? switch switch switch switch PC PC PC PC PC PC PC PC A parallel computer consists
More informationIMARIS At the cutting edge of 3D and 4D interactive analysis and visualization in various application fields
IMARIS At the cutting edge of 3D and 4D interactive analysis and visualization in various application fields Imaris for 3D Visualization, Measurements, Tracking and Interactive Data Analysis Reconstruct
More information