DESIGN AND OVERHEAD ANALYSIS OF WORKFLOWS IN GRID
|
|
- Florence Goodman
- 6 years ago
- Views:
Transcription
1 I J D M S C L Volume 6, o. 1, January-June 2015 DESIG AD OVERHEAD AALYSIS OF WORKFLOWS I GRID S. JAMUA 1, K. REKHA 2, AD R. KAHAVEL 3 ABSRAC Grid workflow execution is approached as a pure best effort scheduling problem that maps the activities onto the Grid processors based on appropriate optimization or local matchmaking heuristics such that the overall execution time is minimized. Even though such heuristics often deliver effective results, the execution in dynamic and unpredictable Grid environments is prone to severe performance losses that must be understood for minimizing the completion time or for the efficient use of high-performance resources. In this paper a new systamtic approach is developed for understanding the sources of the performance losses occurred when executing dynamic workflows in grid environments. Keywords: Distributed system, Grid workflow, Grid performance, Monitoring 1. IRODUCIO Workflows are usually modeled as a graph of activities using graphical composition tools. A scheduler service supported by a Broker or information service is used for optimized use of resources. hese scheduled services are then given to the Grid engine for execution of activities in Grid environments. But this traditional approach does not fully consider the dynamic and course-grain nature of Grid environments such as overheads due to large latencies (several seconds), unpredictable queuing delay, sudden unavailability of existing resources, external loads on Grid resources. Only the pure scheduling strategy is not sufficient for achieving good execution time in Grid environments. We must understand what really happens during the execution of workflows. So performance analysis tools are important for understanding the behaviour and reasons for the performance losses. So that we can improve the run time performance of the Grid environment.. In this paper we propose a systematic approach for developing a performance analysis tool for heterogeneous and distributed grid environment. A theoretical reference parameter called Ideal Execution ime is developed for providing bound for 1,2 Pre final year (Computer Science and Engineering), Government College of Engineering,irunelveli 3 Asst.professor Dept.of Computer Science and Engineering Government College of Engineering,irunelveli
2 64 / S. JAMUA, K. REKHA, AD R. KAHAVEL the lowest execution time in grid environments. hereafter, we present a novel fine-grained classification of non-overlapping workflow overheads covering middleware, data transfer, loss of parallelism, and activity-specific overheads, which contribute to the ideal time to produce the measured Grid execution time of the workflow. We develop a performance tool based on event correlation. Event correlation allows us to reduce the large no of resource specific messages to fewer and more meaningful set of messages. We provide classification of various sets of overheads that occur when executing in Grid environment and validates them using a scalability experiment of a real time application. And we describe a performance analysis tool for analyzing previously described overheads based on event correlation using distributed matrix application. 2. MODEL In this section, we formally define a workflow application based on a restricted directed graph-based structure that represents the result of our experience in modeling and executing several real-world scientific workflows. his model provided us with a good basis for the clear definition of many of the overheads. We model a scientific workflow as a directed graph odes = is the set of activities, (1) C edges = n 1 i [1... n] i ( i, i 1) ( j, k) (2) i 1 j k 1 is the set of control flow dependencies, D Edges ( s, d, D port)` (3) s, d odes is the set of data flow dependencies, An activity can be of the following type, representing workflow regions. 1. Computational activity 2. Parallel region PAR, consisting of set of independent activities. 3. Subworflow We consider the Grid as an aggregation of heterogeneous Grid sites. A Grid site consists of a number of compute and storage systems that share the same local security, network, and resource management policies. Our experimental Grid environments consisting of homogeneous Grid computers within a Grid sites which can the capable of providing their services to remote client computers. Each parallel computers are considered as single computing resources using a local resource management system implemented in Grid Broker computer.
3 DESIG AD OVERHEAD AALYSIS OF WORKFLOWS I GRID / 65 Work Flow activities are typically legacy codes that can be remotely accessed and instantiated using back end application providers, which in our environment called Grid Brokers. Data transfer activities are client systems based on java socket system calls. In this execution model, we consider only the work performed by the workflow computational activities as useful execution time. he time required for any other task is considered a temporal overhead. o measure and compute the temporal overheads, a distributed performance analysis service collects online events generated by sensors distributed across Grid sites, which report on the availability and operation of the middleware services. Figure 1: Example Grid System In our model we use no of clients and M no of resource systems. Resource systems run the dedicated services to client. Client system which wants the service sends their request along with the required data in specified format to resources via a Broker. Broker is responsible for scheduling, monitoring and act as Information System of this Grid model. Broker maintains a dynamic repository which contains information about availability details and binding information of the resource systems.
4 66 / S. JAMUA, K. REKHA, AD R. KAHAVEL We formally define an event generated by distinct sensor as a set of attributes A = {a 1,...a n }. We represent an event as a quadruple: e = <s, t, id, A> where S is the sensor that triggers the event, is the timestamp, id the unique ID of the event, and A is set of private attributes that characterize the event. For example, if a sensor called engine associated with the broker submits at time instance t with the event ID 13, the following event will be raised, e = <engine,t,13, ipadd, out>. On the other hand the following event is created by client to indicate it sends data to broker e = <client, t, eventid, ipadd, {out, broker}> 3. WORKFLOW EMPORAL OVERHEADS o systematically understand the overheads that occur when executing scientific workflows in Grid environments, we model the real execution time w of a workflow W as the sum of an ideal execution time ideal and a set of temporal overheads Ow that originate from various sources: W = ideal + O W (4) Describing the sources of the overheads, classifying them in a detailed hierarchy, and measuring them in a systematic manner are the scope of our analysis. 3.1.Ideal Workflow Execution ime We introduce a so-called ideal execution time that defines a theoretical lower bound for the Grid workflow execution time, assuming the following parameters as ideal premises For our analysis: 1. Zero network latency, 2. Infinite network bandwidth, 3. An optimal workflow schedule, and 4. A sufficiently large number of processors to simultaneously execute all activities of parallel regions (that is, one separate processor per activity). We define the ideal execution time of a computational activity scheduled on a processor P as the fastest wall clock execution time P measured on all processors available on the Grid of the same architecture as P. P = min { } (5) ideal P Grid
5 DESIG AD OVERHEAD AALYSIS OF WORKFLOWS I GRID / 67 he general technique we use to measure P is to submit the executable to the Broker and collect the wall-clock time from the job s standard error stream. We ignore data dependencies in the ideal time since we consider them as an overhead: D-edges = 0. he ideal execution time of a parallel region PAR is the fastest ideal execution time of all its activities: ideal = min { } (6) idel par par par Finally, the ideal execution time of a workflow W is the sum of the ideal execution times of all its activities odes: ideal W = odes (7) ideal We define the total overhead of a workflow W as the difference between its measured Grid execution time W and the ideal time: 3.2. Overhead Classification ideal O w = (8) w w o understand the losses of performance in scientific workflows, we split the total overhead into a hierarchy based on four main categories: middleware, data transfer, loss of parallelism, and activity related Middleware Overhead he middleware overhead is due to the work performed by the middleware services to support the proper execution and completion of the workflow, which we further divide into several components based on the service functionality. Resource Brokerage: his represents the time required by the Resource Broker to query the information service and provide to the Scheduler the processors and activity deployments needed to execute the application. Additionally, this overhead has an important latency component (few seconds), mostly due to the mutual host authentication. his service latency is a common overhead component present in all our middleware services. Performance prediction: his represents the time to provide forecast information about the execution time of individual activities on the Grid sites indicated by the Scheduler, for example, using a polynomial fitting heuristic based on historical or training data. Scheduling: his represents the time to appropriately map the workflow activities onto the Grid resources, which includes the following two sub overheads: 1) a scheduling algorithm, which represents the time required to compute a schedule and 2) rescheduling, which represents the time needed to make a new scheduling decision, for example, because of a performance contract violation or if the workflow changes its runtime structure.
6 68 / S. JAMUA, K. REKHA, AD R. KAHAVEL Execution management: his represents the time to coordinate the execution of workflow activities based on the control and data flow dependencies, which we divide into two categories: Control of parallelism: his represents the time required to fork a set of activities at the beginning of a parallel region or to join them at the end. Job Preparation: his is an overhead that we split into three categories corresponding to the time for (1) partitioning and simplifying the workflow such that multiple activities scheduled on the same site are aggregated into one composite activity and executed as one single job submission (2) archiving, compressing, or uncompressing multiple files that have to be transferred between data-dependent activities or partitions scheduled on different sites to reduce latencies and increase bandwidth utilization; and (3) creating remote directory structures required to execute activities that wrap existing peculiar legacy applications Data ransfer Overhead he data transfer overhead is generated by any kind of data transfer including input/ output file staging between the local computer and the remote Grid site, database access (for example, to store and access historical execution data),third-party file transfer upon large data dependencies, user input (that is, when user interaction is required), imbalanced parallel data transfer (for example, due to different data sizes or network bandwidth), and lost transfer upon workflow rescheduling or rollback. Presently, all these overheads include any network traffic or interruptions since, unlike processors; the wide area network is a shared resource over which we have no control using our current sensors Loss of Parallelism Overhead he loss of parallelism overhead is introduced by parallel regions and has various causes. Serialized Block: his occurs when, as a result of a Scheduling decision, n independent activities (for example, part of a parallel region) are scheduled onto the same Grid site consisting of p processors, where p < n. In this case, the Scheduler introduces the so-called runtime schedule dependencies that prohibit two activities from running simultaneously on the same processor. We call the set of activities of a parallel region PAR serialized through runtime schedule dependencies as a serialized block: B = {1, n par S( 1 ) =... = S(n)}
7 DESIG AD OVERHEAD AALYSIS OF WORKFLOWS I GRID / 69 Where S() denotes the schedule of activity. he execution time of a serialized block is the sum of the serialized activities: B = (9) B For example, assume a parallel region consisting of seven parallel activities in Figure 2. Figure 2: Parallel Region Overhead PAR = 7 i (10) i 1 such that S(1) =S(4) = S(7) = P1, S(2) = S(5) = P2, S(3) = S(6) = P3, and S(3) = S(6) = P3 (where P1 and P2 are part of a dual shared-memory computer). Since the number of parallel activities exceeds the number of processors available, the Scheduler generates three serialized blocks by enhancing the Workflow with the following runtime schedule dependencies : {(1, 4), (4, 7), (2, 5), (3, 6)}. Load imbalance: his occurs in the context of a parallel region when some computational activities finish faster than others and leave the allocated processors idle. We define the load imbalance overhead of a parallel region PAR as the difference between the maximum serialization block end time and the theoretical ideal execution time of the parallel region on the set of processors. Replicated work: his is a technique often applied when performing the same computation redundantly on multiple sites in parallel is faster than computing it on the fastest Grid site and then broadcasting the data. Since replicated jobs are never synchronized and therefore do not produce load imbalance, we calculate the replicated job overhead of a set of identical activities PAR as the ideal time for computing the
8 70 / S. JAMUA, K. REKHA, AD R. KAHAVEL entire replicated work (that is, the harmonic mean of the execution times of all replicated jobs divided by the number of jobs) minus the end time of the fastest job presenting the only useful computation: Activity Overhead 1 ORJ PAR = 1 PAR min { } PAR he activity overhead comprises all overheads that are internal to individual activities, for example, those that we extensively addressed in the previous performance analysis research for parallel programs (for example, communication, load imbalance, cache misses, and synchronization). In addition, we define two new Grid-relevant overheads: 1. he external load of a computational activity on a processor P, which we compute as the difference between the current measured wall-clock time and its ideal execution time: (11) P ideal OEL = (12) 2. lost computation, an important overhead that commonly occurs due to their execution of any computational activity (for example, in case of failures, rollback to the last checkpoint, or when new and substantially more powerful Grid sites become available), which we define as the time difference between the earliest estarted and the latest efailed by the Broker Region-Based Overheads Since every workflow activity produces its own set of overheads, an obvious question is how to aggregate these overheads from individual activities up to parallel regions and the overall workflow level. Let O denotes an arbitrary overhead of an activity. We define the same temporal overhead for an enclosing parallel region O PAR as the sum of the contributed overheads Ocon of all individual activities PAR. he contributed overhead of an activity to an enclosing parallel region is the temporal overhead O averaged across the number of processors used in executing the parallel region. (hat is, as indicated by the Scheduler, which is not necessarily the full available Grid), where each processor P is weighted with its relative speed (execution time of on processor P): O = O, where O PAR PAR CO CO O P Grid And P denotes the wall-clock execution time of activity on processor P. P (13)
9 DESIG AD OVERHEAD AALYSIS OF WORKFLOWS I GRID / 71 Similar to the workflow ideal execution time, we compute the value of any overhead such as middleware, load imbalance, serialization, data transfer, external load, or replicated work for any workflow region by summing the overhead values computed for the underlying computational activities, parallel regions, and sub workflows Unidentified Overheads 0 = odes O (14) Since we carefully designed the overheads as non-overlapping, we can compute the total identified overhead O identified w as the sum of all measured overheads for the entire workflow. We call the difference between the theoretical total overhead O W, and the total unidentified overhead. unidentified identified O W = OW O W (15) Minimizing the unidentified overhead is one important goal of our analysis effort. A high unidentified overhead value indicates that the analysis is unsatisfactory, and further efforts are required to identify new sources of overhead in the workflow execution ormalized Metrics ormalized metrics are used for quantifying the overhead importance with respect to entire workflow execution. We define the severity of a workflow O as the value of the overhead with respect to Grid execution time. SEV O = O W We define the workflow Speed up as the ratio between fastest single-site workflow execution time M Seq and actual Grid execution time W S = min { M } seq M Grid W Further, we define the workflow efficiency as the speedup normalized against the number of Grid sites used, where each site M is weighted with the speed up of the corresponding single-site execution time: (16) (17) S E = M Grid he efficiency formula therefore becomes, S M, where : S M,' M Grid M seq, M seq min { } (18) E = 1 W M ( ) M Grid seq 1 (19) he fastest Grid site has a weight of one, whereas the slowest Grid site has the smallest weight closest to zero.
10 72 / S. JAMUA, K. REKHA, AD R. KAHAVEL 5. COCLUSIO In this paper, we presented a analysis model consisting of a theoretical execution time and the detailed hierarchy of overhead to help the application developers to understand the various sources of bottlenecks that affects the execution of workflows in heterogeneous grid execution model. he purpose of the work is to adjusted well known normalized metrics from parallel processing to the Grid computing scope, including overhead severity, speed up, and efficiency, which are invaluable parameters to be considered before scheduling when the efficient use of resources is an important issue. In the future we are planned to develop solutions for overcome this overheads which affects the execution time of the application. REFERECES [1] Radu Prodan and homas Fahringer, Overhead Analysis of Scientific Workflows in Grid Environments. IEEE rans. on Parallel and Distributed Systems, 19, o. 3, March 2008 [2] K. Czajkowski et al., Grid Information Services for Distributed Resource Sharing, Proc. 10th IEEE Int l Symp. High Performance Distributed Computing (HPDC), [3] R. Vaarandi, SEC-A Lightweight Event Correlation ool, Proc.Workshop IP Operations and Management (IPOM), [4] JAVA Complete Reference by Herbert Schildt.
Assignment 5. Georgia Koloniari
Assignment 5 Georgia Koloniari 2. "Peer-to-Peer Computing" 1. What is the definition of a p2p system given by the authors in sec 1? Compare it with at least one of the definitions surveyed in the last
More informationADAPTIVE AND DYNAMIC LOAD BALANCING METHODOLOGIES FOR DISTRIBUTED ENVIRONMENT
ADAPTIVE AND DYNAMIC LOAD BALANCING METHODOLOGIES FOR DISTRIBUTED ENVIRONMENT PhD Summary DOCTORATE OF PHILOSOPHY IN COMPUTER SCIENCE & ENGINEERING By Sandip Kumar Goyal (09-PhD-052) Under the Supervision
More informationGrid Computing Systems: A Survey and Taxonomy
Grid Computing Systems: A Survey and Taxonomy Material for this lecture from: A Survey and Taxonomy of Resource Management Systems for Grid Computing Systems, K. Krauter, R. Buyya, M. Maheswaran, CS Technical
More informationDISTRIBUTED SYSTEMS Principles and Paradigms Second Edition ANDREW S. TANENBAUM MAARTEN VAN STEEN. Chapter 1. Introduction
DISTRIBUTED SYSTEMS Principles and Paradigms Second Edition ANDREW S. TANENBAUM MAARTEN VAN STEEN Chapter 1 Introduction Modified by: Dr. Ramzi Saifan Definition of a Distributed System (1) A distributed
More informationAdaptive Cluster Computing using JavaSpaces
Adaptive Cluster Computing using JavaSpaces Jyoti Batheja and Manish Parashar The Applied Software Systems Lab. ECE Department, Rutgers University Outline Background Introduction Related Work Summary of
More informationAnalytical Modeling of Parallel Programs
Analytical Modeling of Parallel Programs Alexandre David Introduction to Parallel Computing 1 Topic overview Sources of overhead in parallel programs. Performance metrics for parallel systems. Effect of
More informationOVERHEADS ENHANCEMENT IN MUTIPLE PROCESSING SYSTEMS BY ANURAG REDDY GANKAT KARTHIK REDDY AKKATI
CMPE 655- MULTIPLE PROCESSOR SYSTEMS OVERHEADS ENHANCEMENT IN MUTIPLE PROCESSING SYSTEMS BY ANURAG REDDY GANKAT KARTHIK REDDY AKKATI What is MULTI PROCESSING?? Multiprocessing is the coordinated processing
More informationA RESOURCE MANAGEMENT FRAMEWORK FOR INTERACTIVE GRIDS
A RESOURCE MANAGEMENT FRAMEWORK FOR INTERACTIVE GRIDS Raj Kumar, Vanish Talwar, Sujoy Basu Hewlett-Packard Labs 1501 Page Mill Road, MS 1181 Palo Alto, CA 94304 USA { raj.kumar,vanish.talwar,sujoy.basu}@hp.com
More informationComplexity results for throughput and latency optimization of replicated and data-parallel workflows
Complexity results for throughput and latency optimization of replicated and data-parallel workflows Anne Benoit and Yves Robert GRAAL team, LIP École Normale Supérieure de Lyon June 2007 Anne.Benoit@ens-lyon.fr
More informationParallelization Strategy
COSC 6374 Parallel Computation Algorithm structure Spring 2008 Parallelization Strategy Finding Concurrency Structure the problem to expose exploitable concurrency Algorithm Structure Supporting Structure
More informationIntroduction to parallel computers and parallel programming. Introduction to parallel computersand parallel programming p. 1
Introduction to parallel computers and parallel programming Introduction to parallel computersand parallel programming p. 1 Content A quick overview of morden parallel hardware Parallelism within a chip
More informationLecture 9: Load Balancing & Resource Allocation
Lecture 9: Load Balancing & Resource Allocation Introduction Moler s law, Sullivan s theorem give upper bounds on the speed-up that can be achieved using multiple processors. But to get these need to efficiently
More informationBoundary control : Access Controls: An access control mechanism processes users request for resources in three steps: Identification:
Application control : Boundary control : Access Controls: These controls restrict use of computer system resources to authorized users, limit the actions authorized users can taker with these resources,
More informationOptimize HPC - Application Efficiency on Many Core Systems
Meet the experts Optimize HPC - Application Efficiency on Many Core Systems 2018 Arm Limited Florent Lebeau 27 March 2018 2 2018 Arm Limited Speedup Multithreading and scalability I wrote my program to
More informationChapter 3. Design of Grid Scheduler. 3.1 Introduction
Chapter 3 Design of Grid Scheduler The scheduler component of the grid is responsible to prepare the job ques for grid resources. The research in design of grid schedulers has given various topologies
More informationA Level-wise Priority Based Task Scheduling for Heterogeneous Systems
International Journal of Information and Education Technology, Vol., No. 5, December A Level-wise Priority Based Task Scheduling for Heterogeneous Systems R. Eswari and S. Nickolas, Member IACSIT Abstract
More informationCSCE 626 Experimental Evaluation.
CSCE 626 Experimental Evaluation http://parasol.tamu.edu Introduction This lecture discusses how to properly design an experimental setup, measure and analyze the performance of parallel algorithms you
More information16/10/2008. Today s menu. PRAM Algorithms. What do you program? What do you expect from a model? Basic ideas for the machine model
Today s menu 1. What do you program? Parallel complexity and algorithms PRAM Algorithms 2. The PRAM Model Definition Metrics and notations Brent s principle A few simple algorithms & concepts» Parallel
More informationData Model Considerations for Radar Systems
WHITEPAPER Data Model Considerations for Radar Systems Executive Summary The market demands that today s radar systems be designed to keep up with a rapidly changing threat environment, adapt to new technologies,
More informationNon-uniform memory access machine or (NUMA) is a system where the memory access time to any region of memory is not the same for all processors.
CS 320 Ch. 17 Parallel Processing Multiple Processor Organization The author makes the statement: "Processors execute programs by executing machine instructions in a sequence one at a time." He also says
More informationIBM Spectrum Protect Version Introduction to Data Protection Solutions IBM
IBM Spectrum Protect Version 8.1.2 Introduction to Data Protection Solutions IBM IBM Spectrum Protect Version 8.1.2 Introduction to Data Protection Solutions IBM Note: Before you use this information
More informationA Data-Aware Resource Broker for Data Grids
A Data-Aware Resource Broker for Data Grids Huy Le, Paul Coddington, and Andrew L. Wendelborn School of Computer Science, University of Adelaide Adelaide, SA 5005, Australia {paulc,andrew}@cs.adelaide.edu.au
More informationWhite Paper. Major Performance Tuning Considerations for Weblogic Server
White Paper Major Performance Tuning Considerations for Weblogic Server Table of Contents Introduction and Background Information... 2 Understanding the Performance Objectives... 3 Measuring your Performance
More informationMultimedia Systems 2011/2012
Multimedia Systems 2011/2012 System Architecture Prof. Dr. Paul Müller University of Kaiserslautern Department of Computer Science Integrated Communication Systems ICSY http://www.icsy.de Sitemap 2 Hardware
More informationECE 669 Parallel Computer Architecture
ECE 669 Parallel Computer Architecture Lecture 9 Workload Evaluation Outline Evaluation of applications is important Simulation of sample data sets provides important information Working sets indicate
More informationIBM Tivoli Storage Manager Version Introduction to Data Protection Solutions IBM
IBM Tivoli Storage Manager Version 7.1.6 Introduction to Data Protection Solutions IBM IBM Tivoli Storage Manager Version 7.1.6 Introduction to Data Protection Solutions IBM Note: Before you use this
More informationCSC630/CSC730 Parallel & Distributed Computing
CSC630/CSC730 Parallel & Distributed Computing Analytical Modeling of Parallel Programs Chapter 5 1 Contents Sources of Parallel Overhead Performance Metrics Granularity and Data Mapping Scalability 2
More informationComputer Architecture Lecture 27: Multiprocessors. Prof. Onur Mutlu Carnegie Mellon University Spring 2015, 4/6/2015
18-447 Computer Architecture Lecture 27: Multiprocessors Prof. Onur Mutlu Carnegie Mellon University Spring 2015, 4/6/2015 Assignments Lab 7 out Due April 17 HW 6 Due Friday (April 10) Midterm II April
More informationChapter 13 Strong Scaling
Chapter 13 Strong Scaling Part I. Preliminaries Part II. Tightly Coupled Multicore Chapter 6. Parallel Loops Chapter 7. Parallel Loop Schedules Chapter 8. Parallel Reduction Chapter 9. Reduction Variables
More informationComplexity Results for Throughput and Latency Optimization of Replicated and Data-parallel Workflows
Complexity Results for Throughput and Latency Optimization of Replicated and Data-parallel Workflows Anne Benoit and Yves Robert GRAAL team, LIP École Normale Supérieure de Lyon September 2007 Anne.Benoit@ens-lyon.fr
More informationAnalytical Modeling of Parallel Systems. To accompany the text ``Introduction to Parallel Computing'', Addison Wesley, 2003.
Analytical Modeling of Parallel Systems To accompany the text ``Introduction to Parallel Computing'', Addison Wesley, 2003. Topic Overview Sources of Overhead in Parallel Programs Performance Metrics for
More informationHEP replica management
Primary actor Goal in context Scope Level Stakeholders and interests Precondition Minimal guarantees Success guarantees Trigger Technology and data variations Priority Releases Response time Frequency
More informationLayer-Based Scheduling Algorithms for Multiprocessor-Tasks with Precedence Constraints
Layer-Based Scheduling Algorithms for Multiprocessor-Tasks with Precedence Constraints Jörg Dümmler, Raphael Kunis, and Gudula Rünger Chemnitz University of Technology, Department of Computer Science,
More informationA Framework for Space and Time Efficient Scheduling of Parallelism
A Framework for Space and Time Efficient Scheduling of Parallelism Girija J. Narlikar Guy E. Blelloch December 996 CMU-CS-96-97 School of Computer Science Carnegie Mellon University Pittsburgh, PA 523
More informationL3.4. Data Management Techniques. Frederic Desprez Benjamin Isnard Johan Montagnat
Grid Workflow Efficient Enactment for Data Intensive Applications L3.4 Data Management Techniques Authors : Eddy Caron Frederic Desprez Benjamin Isnard Johan Montagnat Summary : This document presents
More informationAbstract of the Book
Book Keywords IEEE 802.16, IEEE 802.16m, mobile WiMAX, 4G, IMT-Advanced, 3GPP LTE, 3GPP LTE-Advanced, Broadband Wireless, Wireless Communications, Cellular Systems, Network Architecture Abstract of the
More informationClient Server & Distributed System. A Basic Introduction
Client Server & Distributed System A Basic Introduction 1 Client Server Architecture A network architecture in which each computer or process on the network is either a client or a server. Source: http://webopedia.lycos.com
More informationOn Cluster Resource Allocation for Multiple Parallel Task Graphs
On Cluster Resource Allocation for Multiple Parallel Task Graphs Henri Casanova Frédéric Desprez Frédéric Suter University of Hawai i at Manoa INRIA - LIP - ENS Lyon IN2P3 Computing Center, CNRS / IN2P3
More informationOracle Enterprise Manager. 1 Before You Install. System Monitoring Plug-in for Oracle Unified Directory User's Guide Release 1.0
Oracle Enterprise Manager System Monitoring Plug-in for Oracle Unified Directory User's Guide Release 1.0 E24476-01 October 2011 The System Monitoring Plug-In for Oracle Unified Directory extends Oracle
More informationIBM Best Practices Working With Multiple CCM Applications Draft
Best Practices Working With Multiple CCM Applications. This document collects best practices to work with Multiple CCM applications in large size enterprise deployment topologies. Please see Best Practices
More informationUnderstanding Parallelism and the Limitations of Parallel Computing
Understanding Parallelism and the Limitations of Parallel omputing Understanding Parallelism: Sequential work After 16 time steps: 4 cars Scalability Laws 2 Understanding Parallelism: Parallel work After
More informationEmbedded Systems Dr. Santanu Chaudhury Department of Electrical Engineering Indian Institution of Technology, Delhi
Embedded Systems Dr. Santanu Chaudhury Department of Electrical Engineering Indian Institution of Technology, Delhi Lecture - 34 Compilers for Embedded Systems Today, we shall look at the compilers, which
More informationBest Practices. Deploying Optim Performance Manager in large scale environments. IBM Optim Performance Manager Extended Edition V4.1.0.
IBM Optim Performance Manager Extended Edition V4.1.0.1 Best Practices Deploying Optim Performance Manager in large scale environments Ute Baumbach (bmb@de.ibm.com) Optim Performance Manager Development
More informationKismet: Parallel Speedup Estimates for Serial Programs
Kismet: Parallel Speedup Estimates for Serial Programs Donghwan Jeon, Saturnino Garcia, Chris Louie, and Michael Bedford Taylor Computer Science and Engineering University of California, San Diego 1 Questions
More informationHigh Performance Computing in C and C++
High Performance Computing in C and C++ Rita Borgo Computer Science Department, Swansea University Announcement No change in lecture schedule: Timetable remains the same: Monday 1 to 2 Glyndwr C Friday
More informationCSE398: Network Systems Design
CSE398: Network Systems Design Instructor: Dr. Liang Cheng Department of Computer Science and Engineering P.C. Rossin College of Engineering & Applied Science Lehigh University March 14, 2005 Outline Classification
More informationA Top Catching Scheme Consistency Controlling in Hybrid P2P Network
A Top Catching Scheme Consistency Controlling in Hybrid P2P Network V. Asha*1, P Ramesh Babu*2 M.Tech (CSE) Student Department of CSE, Priyadarshini Institute of Technology & Science, Chintalapudi, Guntur(Dist),
More informationGrADSoft and its Application Manager: An Execution Mechanism for Grid Applications
GrADSoft and its Application Manager: An Execution Mechanism for Grid Applications Authors Ken Kennedy, Mark Mazina, John Mellor-Crummey, Rice University Ruth Aydt, Celso Mendes, UIUC Holly Dail, Otto
More informationEfficient, Scalable, and Provenance-Aware Management of Linked Data
Efficient, Scalable, and Provenance-Aware Management of Linked Data Marcin Wylot 1 Motivation and objectives of the research The proliferation of heterogeneous Linked Data on the Web requires data management
More informationSolidFire and Pure Storage Architectural Comparison
The All-Flash Array Built for the Next Generation Data Center SolidFire and Pure Storage Architectural Comparison June 2014 This document includes general information about Pure Storage architecture as
More informationCLOUD-SCALE FILE SYSTEMS
Data Management in the Cloud CLOUD-SCALE FILE SYSTEMS 92 Google File System (GFS) Designing a file system for the Cloud design assumptions design choices Architecture GFS Master GFS Chunkservers GFS Clients
More informationChapter 5: Analytical Modelling of Parallel Programs
Chapter 5: Analytical Modelling of Parallel Programs Introduction to Parallel Computing, Second Edition By Ananth Grama, Anshul Gupta, George Karypis, Vipin Kumar Contents 1. Sources of Overhead in Parallel
More informationGraph Partitioning for Scalable Distributed Graph Computations
Graph Partitioning for Scalable Distributed Graph Computations Aydın Buluç ABuluc@lbl.gov Kamesh Madduri madduri@cse.psu.edu 10 th DIMACS Implementation Challenge, Graph Partitioning and Graph Clustering
More informationTag a Tiny Aggregation Service for Ad-Hoc Sensor Networks. Samuel Madden, Michael Franklin, Joseph Hellerstein,Wei Hong UC Berkeley Usinex OSDI 02
Tag a Tiny Aggregation Service for Ad-Hoc Sensor Networks Samuel Madden, Michael Franklin, Joseph Hellerstein,Wei Hong UC Berkeley Usinex OSDI 02 Outline Introduction The Tiny AGgregation Approach Aggregate
More informationDesigning Issues For Distributed Computing System: An Empirical View
ISSN: 2278 0211 (Online) Designing Issues For Distributed Computing System: An Empirical View Dr. S.K Gandhi, Research Guide Department of Computer Science & Engineering, AISECT University, Bhopal (M.P),
More informationDesign of Parallel Algorithms. Models of Parallel Computation
+ Design of Parallel Algorithms Models of Parallel Computation + Chapter Overview: Algorithms and Concurrency n Introduction to Parallel Algorithms n Tasks and Decomposition n Processes and Mapping n Processes
More informationParallel DBMS. Parallel Database Systems. PDBS vs Distributed DBS. Types of Parallelism. Goals and Metrics Speedup. Types of Parallelism
Parallel DBMS Parallel Database Systems CS5225 Parallel DB 1 Uniprocessor technology has reached its limit Difficult to build machines powerful enough to meet the CPU and I/O demands of DBMS serving large
More informationAdaptive Aggregation Scheduling Using. Aggregation-degree Control in Sensor Network
Contemporary Engineering Sciences, Vol. 7, 2014, no. 14, 725-730 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ces.2014.4681 Adaptive Aggregation Scheduling Using Aggregation-degree Control in
More information1 Dulcian, Inc., 2001 All rights reserved. Oracle9i Data Warehouse Review. Agenda
Agenda Oracle9i Warehouse Review Dulcian, Inc. Oracle9i Server OLAP Server Analytical SQL Mining ETL Infrastructure 9i Warehouse Builder Oracle 9i Server Overview E-Business Intelligence Platform 9i Server:
More informationComplexity Measures for Map-Reduce, and Comparison to Parallel Computing
Complexity Measures for Map-Reduce, and Comparison to Parallel Computing Ashish Goel Stanford University and Twitter Kamesh Munagala Duke University and Twitter November 11, 2012 The programming paradigm
More informationSCHISM: A WORKLOAD-DRIVEN APPROACH TO DATABASE REPLICATION AND PARTITIONING
SCHISM: A WORKLOAD-DRIVEN APPROACH TO DATABASE REPLICATION AND PARTITIONING ZEYNEP KORKMAZ CS742 - PARALLEL AND DISTRIBUTED DATABASE SYSTEMS UNIVERSITY OF WATERLOO OUTLINE. Background 2. What is Schism?
More informationAnalyzing the Performance of IWAVE on a Cluster using HPCToolkit
Analyzing the Performance of IWAVE on a Cluster using HPCToolkit John Mellor-Crummey and Laksono Adhianto Department of Computer Science Rice University {johnmc,laksono}@rice.edu TRIP Meeting March 30,
More informationA Partial Critical Path Based Approach for Grid Workflow Scheduling
A Partial Critical Path Based Approach for Grid Workflow Scheduling Anagha Sharaf 1, Suguna.M 2 PG Scholar, Department of IT, S.N.S College of Technology, Coimbatore, Tamilnadu, India 1 Associate Professor,
More informationPerformance Extrapolation for Load Testing Results of Mixture of Applications
Performance Extrapolation for Load Testing Results of Mixture of Applications Subhasri Duttagupta, Manoj Nambiar Tata Innovation Labs, Performance Engineering Research Center Tata Consulting Services Mumbai,
More informationGrid Architectural Models
Grid Architectural Models Computational Grids - A computational Grid aggregates the processing power from a distributed collection of systems - This type of Grid is primarily composed of low powered computers
More informationGrid Scheduler. Grid Information Service. Local Resource Manager L l Resource Manager. Single CPU (Time Shared Allocation) (Space Shared Allocation)
Scheduling on the Grid 1 2 Grid Scheduling Architecture User Application Grid Scheduler Grid Information Service Local Resource Manager Local Resource Manager Local L l Resource Manager 2100 2100 2100
More informationAn Introduction to Parallel Programming
An Introduction to Parallel Programming Ing. Andrea Marongiu (a.marongiu@unibo.it) Includes slides from Multicore Programming Primer course at Massachusetts Institute of Technology (MIT) by Prof. SamanAmarasinghe
More informationTensorFlow: A System for Learning-Scale Machine Learning. Google Brain
TensorFlow: A System for Learning-Scale Machine Learning Google Brain The Problem Machine learning is everywhere This is in large part due to: 1. Invention of more sophisticated machine learning models
More informationLecture 7: Parallel Processing
Lecture 7: Parallel Processing Introduction and motivation Architecture classification Performance evaluation Interconnection network Zebo Peng, IDA, LiTH 1 Performance Improvement Reduction of instruction
More informationMitigating Data Skew Using Map Reduce Application
Ms. Archana P.M Mitigating Data Skew Using Map Reduce Application Mr. Malathesh S.H 4 th sem, M.Tech (C.S.E) Associate Professor C.S.E Dept. M.S.E.C, V.T.U Bangalore, India archanaanil062@gmail.com M.S.E.C,
More informationIssues in Parallel Processing. Lecture for CPSC 5155 Edward Bosworth, Ph.D. Computer Science Department Columbus State University
Issues in Parallel Processing Lecture for CPSC 5155 Edward Bosworth, Ph.D. Computer Science Department Columbus State University Introduction Goal: connecting multiple computers to get higher performance
More informationA Survey on Grid Scheduling Systems
Technical Report Report #: SJTU_CS_TR_200309001 A Survey on Grid Scheduling Systems Yanmin Zhu and Lionel M Ni Cite this paper: Yanmin Zhu, Lionel M. Ni, A Survey on Grid Scheduling Systems, Technical
More informationEnterprise print management in VMware Horizon
Enterprise print management in VMware Horizon Introduction: Embracing and Extending VMware Horizon Tricerat Simplify Printing enhances the capabilities of VMware Horizon environments by enabling reliable
More informationOverview SENTINET 3.1
Overview SENTINET 3.1 Overview 1 Contents Introduction... 2 Customer Benefits... 3 Development and Test... 3 Production and Operations... 4 Architecture... 5 Technology Stack... 7 Features Summary... 7
More informationLoad Balancing for Parallel Multi-core Machines with Non-Uniform Communication Costs
Load Balancing for Parallel Multi-core Machines with Non-Uniform Communication Costs Laércio Lima Pilla llpilla@inf.ufrgs.br LIG Laboratory INRIA Grenoble University Grenoble, France Institute of Informatics
More informationMicrosoft SQL Server Fix Pack 15. Reference IBM
Microsoft SQL Server 6.3.1 Fix Pack 15 Reference IBM Microsoft SQL Server 6.3.1 Fix Pack 15 Reference IBM Note Before using this information and the product it supports, read the information in Notices
More informationScalable Computing: Practice and Experience Volume 10, Number 4, pp
Scalable Computing: Practice and Experience Volume 10, Number 4, pp. 413 418. http://www.scpe.org ISSN 1895-1767 c 2009 SCPE MULTI-APPLICATION BAG OF JOBS FOR INTERACTIVE AND ON-DEMAND COMPUTING BRANKO
More informationTemporal Exception Prediction for Loops in Resource Constrained Concurrent Workflows
emporal Exception Prediction for Loops in Resource Constrained Concurrent Workflows Iok-Fai Leong, Yain-Whar Si Faculty of Science and echnology, University of Macau {henryifl, fstasp}@umac.mo Abstract
More informationCompressing and Decoding Term Statistics Time Series
Compressing and Decoding Term Statistics Time Series Jinfeng Rao 1,XingNiu 1,andJimmyLin 2(B) 1 University of Maryland, College Park, USA {jinfeng,xingniu}@cs.umd.edu 2 University of Waterloo, Waterloo,
More informationJAVA IEEE TRANSACTION ON CLOUD COMPUTING. 1. ITJCC01 Nebula: Distributed Edge Cloud for Data Intensive Computing
JAVA IEEE TRANSACTION ON CLOUD COMPUTING 1. ITJCC01 Nebula: Distributed Edge for Data Intensive Computing 2. ITJCC02 A semi-automatic and trustworthy scheme for continuous cloud service certification 3.
More informationUNIT I (Two Marks Questions & Answers)
UNIT I (Two Marks Questions & Answers) Discuss the different ways how instruction set architecture can be classified? Stack Architecture,Accumulator Architecture, Register-Memory Architecture,Register-
More informationDISTRIBUTED HIGH-SPEED COMPUTING OF MULTIMEDIA DATA
DISTRIBUTED HIGH-SPEED COMPUTING OF MULTIMEDIA DATA M. GAUS, G. R. JOUBERT, O. KAO, S. RIEDEL AND S. STAPEL Technical University of Clausthal, Department of Computer Science Julius-Albert-Str. 4, 38678
More informationCorral: A Glide-in Based Service for Resource Provisioning
: A Glide-in Based Service for Resource Provisioning Gideon Juve USC Information Sciences Institute juve@usc.edu Outline Throughput Applications Grid Computing Multi-level scheduling and Glideins Example:
More informationParallel Programming Concepts. Parallel Algorithms. Peter Tröger
Parallel Programming Concepts Parallel Algorithms Peter Tröger Sources: Ian Foster. Designing and Building Parallel Programs. Addison-Wesley. 1995. Mattson, Timothy G.; S, Beverly A.; ers,; Massingill,
More informationA Taxonomy and Survey of Grid Resource Management Systems
A Taxonomy and Survey of Grid Resource Management Systems Klaus Krauter 1, Rajkumar Buyya 2, and Muthucumaru Maheswaran 1 Advanced Networking Research Laboratory 1 Department of Computer Science University
More informationMapping a group of jobs in the error recovery of the Grid-based workflow within SLA context
Mapping a group of jobs in the error recovery of the Grid-based workflow within SLA context Dang Minh Quan International University in Germany School of Information Technology Bruchsal 76646, Germany quandm@upb.de
More informationExample: CPU-bound process that would run for 100 quanta continuously 1, 2, 4, 8, 16, 32, 64 (only 37 required for last run) Needs only 7 swaps
Interactive Scheduling Algorithms Continued o Priority Scheduling Introduction Round-robin assumes all processes are equal often not the case Assign a priority to each process, and always choose the process
More informationA Study of High Performance Computing and the Cray SV1 Supercomputer. Michael Sullivan TJHSST Class of 2004
A Study of High Performance Computing and the Cray SV1 Supercomputer Michael Sullivan TJHSST Class of 2004 June 2004 0.1 Introduction A supercomputer is a device for turning compute-bound problems into
More informationKenneth A. Hawick P. D. Coddington H. A. James
Student: Vidar Tulinius Email: vidarot@brandeis.edu Distributed frameworks and parallel algorithms for processing large-scale geographic data Kenneth A. Hawick P. D. Coddington H. A. James Overview Increasing
More informationIntroduction to Distributed Systems. INF5040/9040 Autumn 2018 Lecturer: Eli Gjørven (ifi/uio)
Introduction to Distributed Systems INF5040/9040 Autumn 2018 Lecturer: Eli Gjørven (ifi/uio) August 28, 2018 Outline Definition of a distributed system Goals of a distributed system Implications of distributed
More informationCS6030 Cloud Computing. Acknowledgements. Today s Topics. Intro to Cloud Computing 10/20/15. Ajay Gupta, WMU-CS. WiSe Lab
CS6030 Cloud Computing Ajay Gupta B239, CEAS Computer Science Department Western Michigan University ajay.gupta@wmich.edu 276-3104 1 Acknowledgements I have liberally borrowed these slides and material
More informationSolution of Out-of-Core Lower-Upper Decomposition for Complex Valued Matrices
Solution of Out-of-Core Lower-Upper Decomposition for Complex Valued Matrices Marianne Spurrier and Joe Swartz, Lockheed Martin Corp. and ruce lack, Cray Inc. ASTRACT: Matrix decomposition and solution
More informationA Scalable Content- Addressable Network
A Scalable Content- Addressable Network In Proceedings of ACM SIGCOMM 2001 S. Ratnasamy, P. Francis, M. Handley, R. Karp, S. Shenker Presented by L.G. Alex Sung 9th March 2005 for CS856 1 Outline CAN basics
More informationTHE GLOBUS PROJECT. White Paper. GridFTP. Universal Data Transfer for the Grid
THE GLOBUS PROJECT White Paper GridFTP Universal Data Transfer for the Grid WHITE PAPER GridFTP Universal Data Transfer for the Grid September 5, 2000 Copyright 2000, The University of Chicago and The
More informationMemory Management Algorithms on Distributed Systems. Katie Becker and David Rodgers CS425 April 15, 2005
Memory Management Algorithms on Distributed Systems Katie Becker and David Rodgers CS425 April 15, 2005 Table of Contents 1. Introduction 2. Coarse Grained Memory 2.1. Bottlenecks 2.2. Simulations 2.3.
More informationVerification and Validation of X-Sim: A Trace-Based Simulator
http://www.cse.wustl.edu/~jain/cse567-06/ftp/xsim/index.html 1 of 11 Verification and Validation of X-Sim: A Trace-Based Simulator Saurabh Gayen, sg3@wustl.edu Abstract X-Sim is a trace-based simulator
More informationAuto Management for Apache Kafka and Distributed Stateful System in General
Auto Management for Apache Kafka and Distributed Stateful System in General Jiangjie (Becket) Qin Data Infrastructure @LinkedIn GIAC 2017, 12/23/17@Shanghai Agenda Kafka introduction and terminologies
More informationImproving the Efficiency of Fast Using Semantic Similarity Algorithm
International Journal of Scientific and Research Publications, Volume 4, Issue 1, January 2014 1 Improving the Efficiency of Fast Using Semantic Similarity Algorithm D.KARTHIKA 1, S. DIVAKAR 2 Final year
More informationAnna Morajko.
Performance analysis and tuning of parallel/distributed applications Anna Morajko Anna.Morajko@uab.es 26 05 2008 Introduction Main research projects Develop techniques and tools for application performance
More informationPrinciples of Parallel Algorithm Design: Concurrency and Mapping
Principles of Parallel Algorithm Design: Concurrency and Mapping John Mellor-Crummey Department of Computer Science Rice University johnmc@rice.edu COMP 422/534 Lecture 3 17 January 2017 Last Thursday
More information