DESIGN AND OVERHEAD ANALYSIS OF WORKFLOWS IN GRID

Size: px
Start display at page:

Download "DESIGN AND OVERHEAD ANALYSIS OF WORKFLOWS IN GRID"

Transcription

1 I J D M S C L Volume 6, o. 1, January-June 2015 DESIG AD OVERHEAD AALYSIS OF WORKFLOWS I GRID S. JAMUA 1, K. REKHA 2, AD R. KAHAVEL 3 ABSRAC Grid workflow execution is approached as a pure best effort scheduling problem that maps the activities onto the Grid processors based on appropriate optimization or local matchmaking heuristics such that the overall execution time is minimized. Even though such heuristics often deliver effective results, the execution in dynamic and unpredictable Grid environments is prone to severe performance losses that must be understood for minimizing the completion time or for the efficient use of high-performance resources. In this paper a new systamtic approach is developed for understanding the sources of the performance losses occurred when executing dynamic workflows in grid environments. Keywords: Distributed system, Grid workflow, Grid performance, Monitoring 1. IRODUCIO Workflows are usually modeled as a graph of activities using graphical composition tools. A scheduler service supported by a Broker or information service is used for optimized use of resources. hese scheduled services are then given to the Grid engine for execution of activities in Grid environments. But this traditional approach does not fully consider the dynamic and course-grain nature of Grid environments such as overheads due to large latencies (several seconds), unpredictable queuing delay, sudden unavailability of existing resources, external loads on Grid resources. Only the pure scheduling strategy is not sufficient for achieving good execution time in Grid environments. We must understand what really happens during the execution of workflows. So performance analysis tools are important for understanding the behaviour and reasons for the performance losses. So that we can improve the run time performance of the Grid environment.. In this paper we propose a systematic approach for developing a performance analysis tool for heterogeneous and distributed grid environment. A theoretical reference parameter called Ideal Execution ime is developed for providing bound for 1,2 Pre final year (Computer Science and Engineering), Government College of Engineering,irunelveli 3 Asst.professor Dept.of Computer Science and Engineering Government College of Engineering,irunelveli

2 64 / S. JAMUA, K. REKHA, AD R. KAHAVEL the lowest execution time in grid environments. hereafter, we present a novel fine-grained classification of non-overlapping workflow overheads covering middleware, data transfer, loss of parallelism, and activity-specific overheads, which contribute to the ideal time to produce the measured Grid execution time of the workflow. We develop a performance tool based on event correlation. Event correlation allows us to reduce the large no of resource specific messages to fewer and more meaningful set of messages. We provide classification of various sets of overheads that occur when executing in Grid environment and validates them using a scalability experiment of a real time application. And we describe a performance analysis tool for analyzing previously described overheads based on event correlation using distributed matrix application. 2. MODEL In this section, we formally define a workflow application based on a restricted directed graph-based structure that represents the result of our experience in modeling and executing several real-world scientific workflows. his model provided us with a good basis for the clear definition of many of the overheads. We model a scientific workflow as a directed graph odes = is the set of activities, (1) C edges = n 1 i [1... n] i ( i, i 1) ( j, k) (2) i 1 j k 1 is the set of control flow dependencies, D Edges ( s, d, D port)` (3) s, d odes is the set of data flow dependencies, An activity can be of the following type, representing workflow regions. 1. Computational activity 2. Parallel region PAR, consisting of set of independent activities. 3. Subworflow We consider the Grid as an aggregation of heterogeneous Grid sites. A Grid site consists of a number of compute and storage systems that share the same local security, network, and resource management policies. Our experimental Grid environments consisting of homogeneous Grid computers within a Grid sites which can the capable of providing their services to remote client computers. Each parallel computers are considered as single computing resources using a local resource management system implemented in Grid Broker computer.

3 DESIG AD OVERHEAD AALYSIS OF WORKFLOWS I GRID / 65 Work Flow activities are typically legacy codes that can be remotely accessed and instantiated using back end application providers, which in our environment called Grid Brokers. Data transfer activities are client systems based on java socket system calls. In this execution model, we consider only the work performed by the workflow computational activities as useful execution time. he time required for any other task is considered a temporal overhead. o measure and compute the temporal overheads, a distributed performance analysis service collects online events generated by sensors distributed across Grid sites, which report on the availability and operation of the middleware services. Figure 1: Example Grid System In our model we use no of clients and M no of resource systems. Resource systems run the dedicated services to client. Client system which wants the service sends their request along with the required data in specified format to resources via a Broker. Broker is responsible for scheduling, monitoring and act as Information System of this Grid model. Broker maintains a dynamic repository which contains information about availability details and binding information of the resource systems.

4 66 / S. JAMUA, K. REKHA, AD R. KAHAVEL We formally define an event generated by distinct sensor as a set of attributes A = {a 1,...a n }. We represent an event as a quadruple: e = <s, t, id, A> where S is the sensor that triggers the event, is the timestamp, id the unique ID of the event, and A is set of private attributes that characterize the event. For example, if a sensor called engine associated with the broker submits at time instance t with the event ID 13, the following event will be raised, e = <engine,t,13, ipadd, out>. On the other hand the following event is created by client to indicate it sends data to broker e = <client, t, eventid, ipadd, {out, broker}> 3. WORKFLOW EMPORAL OVERHEADS o systematically understand the overheads that occur when executing scientific workflows in Grid environments, we model the real execution time w of a workflow W as the sum of an ideal execution time ideal and a set of temporal overheads Ow that originate from various sources: W = ideal + O W (4) Describing the sources of the overheads, classifying them in a detailed hierarchy, and measuring them in a systematic manner are the scope of our analysis. 3.1.Ideal Workflow Execution ime We introduce a so-called ideal execution time that defines a theoretical lower bound for the Grid workflow execution time, assuming the following parameters as ideal premises For our analysis: 1. Zero network latency, 2. Infinite network bandwidth, 3. An optimal workflow schedule, and 4. A sufficiently large number of processors to simultaneously execute all activities of parallel regions (that is, one separate processor per activity). We define the ideal execution time of a computational activity scheduled on a processor P as the fastest wall clock execution time P measured on all processors available on the Grid of the same architecture as P. P = min { } (5) ideal P Grid

5 DESIG AD OVERHEAD AALYSIS OF WORKFLOWS I GRID / 67 he general technique we use to measure P is to submit the executable to the Broker and collect the wall-clock time from the job s standard error stream. We ignore data dependencies in the ideal time since we consider them as an overhead: D-edges = 0. he ideal execution time of a parallel region PAR is the fastest ideal execution time of all its activities: ideal = min { } (6) idel par par par Finally, the ideal execution time of a workflow W is the sum of the ideal execution times of all its activities odes: ideal W = odes (7) ideal We define the total overhead of a workflow W as the difference between its measured Grid execution time W and the ideal time: 3.2. Overhead Classification ideal O w = (8) w w o understand the losses of performance in scientific workflows, we split the total overhead into a hierarchy based on four main categories: middleware, data transfer, loss of parallelism, and activity related Middleware Overhead he middleware overhead is due to the work performed by the middleware services to support the proper execution and completion of the workflow, which we further divide into several components based on the service functionality. Resource Brokerage: his represents the time required by the Resource Broker to query the information service and provide to the Scheduler the processors and activity deployments needed to execute the application. Additionally, this overhead has an important latency component (few seconds), mostly due to the mutual host authentication. his service latency is a common overhead component present in all our middleware services. Performance prediction: his represents the time to provide forecast information about the execution time of individual activities on the Grid sites indicated by the Scheduler, for example, using a polynomial fitting heuristic based on historical or training data. Scheduling: his represents the time to appropriately map the workflow activities onto the Grid resources, which includes the following two sub overheads: 1) a scheduling algorithm, which represents the time required to compute a schedule and 2) rescheduling, which represents the time needed to make a new scheduling decision, for example, because of a performance contract violation or if the workflow changes its runtime structure.

6 68 / S. JAMUA, K. REKHA, AD R. KAHAVEL Execution management: his represents the time to coordinate the execution of workflow activities based on the control and data flow dependencies, which we divide into two categories: Control of parallelism: his represents the time required to fork a set of activities at the beginning of a parallel region or to join them at the end. Job Preparation: his is an overhead that we split into three categories corresponding to the time for (1) partitioning and simplifying the workflow such that multiple activities scheduled on the same site are aggregated into one composite activity and executed as one single job submission (2) archiving, compressing, or uncompressing multiple files that have to be transferred between data-dependent activities or partitions scheduled on different sites to reduce latencies and increase bandwidth utilization; and (3) creating remote directory structures required to execute activities that wrap existing peculiar legacy applications Data ransfer Overhead he data transfer overhead is generated by any kind of data transfer including input/ output file staging between the local computer and the remote Grid site, database access (for example, to store and access historical execution data),third-party file transfer upon large data dependencies, user input (that is, when user interaction is required), imbalanced parallel data transfer (for example, due to different data sizes or network bandwidth), and lost transfer upon workflow rescheduling or rollback. Presently, all these overheads include any network traffic or interruptions since, unlike processors; the wide area network is a shared resource over which we have no control using our current sensors Loss of Parallelism Overhead he loss of parallelism overhead is introduced by parallel regions and has various causes. Serialized Block: his occurs when, as a result of a Scheduling decision, n independent activities (for example, part of a parallel region) are scheduled onto the same Grid site consisting of p processors, where p < n. In this case, the Scheduler introduces the so-called runtime schedule dependencies that prohibit two activities from running simultaneously on the same processor. We call the set of activities of a parallel region PAR serialized through runtime schedule dependencies as a serialized block: B = {1, n par S( 1 ) =... = S(n)}

7 DESIG AD OVERHEAD AALYSIS OF WORKFLOWS I GRID / 69 Where S() denotes the schedule of activity. he execution time of a serialized block is the sum of the serialized activities: B = (9) B For example, assume a parallel region consisting of seven parallel activities in Figure 2. Figure 2: Parallel Region Overhead PAR = 7 i (10) i 1 such that S(1) =S(4) = S(7) = P1, S(2) = S(5) = P2, S(3) = S(6) = P3, and S(3) = S(6) = P3 (where P1 and P2 are part of a dual shared-memory computer). Since the number of parallel activities exceeds the number of processors available, the Scheduler generates three serialized blocks by enhancing the Workflow with the following runtime schedule dependencies : {(1, 4), (4, 7), (2, 5), (3, 6)}. Load imbalance: his occurs in the context of a parallel region when some computational activities finish faster than others and leave the allocated processors idle. We define the load imbalance overhead of a parallel region PAR as the difference between the maximum serialization block end time and the theoretical ideal execution time of the parallel region on the set of processors. Replicated work: his is a technique often applied when performing the same computation redundantly on multiple sites in parallel is faster than computing it on the fastest Grid site and then broadcasting the data. Since replicated jobs are never synchronized and therefore do not produce load imbalance, we calculate the replicated job overhead of a set of identical activities PAR as the ideal time for computing the

8 70 / S. JAMUA, K. REKHA, AD R. KAHAVEL entire replicated work (that is, the harmonic mean of the execution times of all replicated jobs divided by the number of jobs) minus the end time of the fastest job presenting the only useful computation: Activity Overhead 1 ORJ PAR = 1 PAR min { } PAR he activity overhead comprises all overheads that are internal to individual activities, for example, those that we extensively addressed in the previous performance analysis research for parallel programs (for example, communication, load imbalance, cache misses, and synchronization). In addition, we define two new Grid-relevant overheads: 1. he external load of a computational activity on a processor P, which we compute as the difference between the current measured wall-clock time and its ideal execution time: (11) P ideal OEL = (12) 2. lost computation, an important overhead that commonly occurs due to their execution of any computational activity (for example, in case of failures, rollback to the last checkpoint, or when new and substantially more powerful Grid sites become available), which we define as the time difference between the earliest estarted and the latest efailed by the Broker Region-Based Overheads Since every workflow activity produces its own set of overheads, an obvious question is how to aggregate these overheads from individual activities up to parallel regions and the overall workflow level. Let O denotes an arbitrary overhead of an activity. We define the same temporal overhead for an enclosing parallel region O PAR as the sum of the contributed overheads Ocon of all individual activities PAR. he contributed overhead of an activity to an enclosing parallel region is the temporal overhead O averaged across the number of processors used in executing the parallel region. (hat is, as indicated by the Scheduler, which is not necessarily the full available Grid), where each processor P is weighted with its relative speed (execution time of on processor P): O = O, where O PAR PAR CO CO O P Grid And P denotes the wall-clock execution time of activity on processor P. P (13)

9 DESIG AD OVERHEAD AALYSIS OF WORKFLOWS I GRID / 71 Similar to the workflow ideal execution time, we compute the value of any overhead such as middleware, load imbalance, serialization, data transfer, external load, or replicated work for any workflow region by summing the overhead values computed for the underlying computational activities, parallel regions, and sub workflows Unidentified Overheads 0 = odes O (14) Since we carefully designed the overheads as non-overlapping, we can compute the total identified overhead O identified w as the sum of all measured overheads for the entire workflow. We call the difference between the theoretical total overhead O W, and the total unidentified overhead. unidentified identified O W = OW O W (15) Minimizing the unidentified overhead is one important goal of our analysis effort. A high unidentified overhead value indicates that the analysis is unsatisfactory, and further efforts are required to identify new sources of overhead in the workflow execution ormalized Metrics ormalized metrics are used for quantifying the overhead importance with respect to entire workflow execution. We define the severity of a workflow O as the value of the overhead with respect to Grid execution time. SEV O = O W We define the workflow Speed up as the ratio between fastest single-site workflow execution time M Seq and actual Grid execution time W S = min { M } seq M Grid W Further, we define the workflow efficiency as the speedup normalized against the number of Grid sites used, where each site M is weighted with the speed up of the corresponding single-site execution time: (16) (17) S E = M Grid he efficiency formula therefore becomes, S M, where : S M,' M Grid M seq, M seq min { } (18) E = 1 W M ( ) M Grid seq 1 (19) he fastest Grid site has a weight of one, whereas the slowest Grid site has the smallest weight closest to zero.

10 72 / S. JAMUA, K. REKHA, AD R. KAHAVEL 5. COCLUSIO In this paper, we presented a analysis model consisting of a theoretical execution time and the detailed hierarchy of overhead to help the application developers to understand the various sources of bottlenecks that affects the execution of workflows in heterogeneous grid execution model. he purpose of the work is to adjusted well known normalized metrics from parallel processing to the Grid computing scope, including overhead severity, speed up, and efficiency, which are invaluable parameters to be considered before scheduling when the efficient use of resources is an important issue. In the future we are planned to develop solutions for overcome this overheads which affects the execution time of the application. REFERECES [1] Radu Prodan and homas Fahringer, Overhead Analysis of Scientific Workflows in Grid Environments. IEEE rans. on Parallel and Distributed Systems, 19, o. 3, March 2008 [2] K. Czajkowski et al., Grid Information Services for Distributed Resource Sharing, Proc. 10th IEEE Int l Symp. High Performance Distributed Computing (HPDC), [3] R. Vaarandi, SEC-A Lightweight Event Correlation ool, Proc.Workshop IP Operations and Management (IPOM), [4] JAVA Complete Reference by Herbert Schildt.

Assignment 5. Georgia Koloniari

Assignment 5. Georgia Koloniari Assignment 5 Georgia Koloniari 2. "Peer-to-Peer Computing" 1. What is the definition of a p2p system given by the authors in sec 1? Compare it with at least one of the definitions surveyed in the last

More information

ADAPTIVE AND DYNAMIC LOAD BALANCING METHODOLOGIES FOR DISTRIBUTED ENVIRONMENT

ADAPTIVE AND DYNAMIC LOAD BALANCING METHODOLOGIES FOR DISTRIBUTED ENVIRONMENT ADAPTIVE AND DYNAMIC LOAD BALANCING METHODOLOGIES FOR DISTRIBUTED ENVIRONMENT PhD Summary DOCTORATE OF PHILOSOPHY IN COMPUTER SCIENCE & ENGINEERING By Sandip Kumar Goyal (09-PhD-052) Under the Supervision

More information

Grid Computing Systems: A Survey and Taxonomy

Grid Computing Systems: A Survey and Taxonomy Grid Computing Systems: A Survey and Taxonomy Material for this lecture from: A Survey and Taxonomy of Resource Management Systems for Grid Computing Systems, K. Krauter, R. Buyya, M. Maheswaran, CS Technical

More information

DISTRIBUTED SYSTEMS Principles and Paradigms Second Edition ANDREW S. TANENBAUM MAARTEN VAN STEEN. Chapter 1. Introduction

DISTRIBUTED SYSTEMS Principles and Paradigms Second Edition ANDREW S. TANENBAUM MAARTEN VAN STEEN. Chapter 1. Introduction DISTRIBUTED SYSTEMS Principles and Paradigms Second Edition ANDREW S. TANENBAUM MAARTEN VAN STEEN Chapter 1 Introduction Modified by: Dr. Ramzi Saifan Definition of a Distributed System (1) A distributed

More information

Adaptive Cluster Computing using JavaSpaces

Adaptive Cluster Computing using JavaSpaces Adaptive Cluster Computing using JavaSpaces Jyoti Batheja and Manish Parashar The Applied Software Systems Lab. ECE Department, Rutgers University Outline Background Introduction Related Work Summary of

More information

Analytical Modeling of Parallel Programs

Analytical Modeling of Parallel Programs Analytical Modeling of Parallel Programs Alexandre David Introduction to Parallel Computing 1 Topic overview Sources of overhead in parallel programs. Performance metrics for parallel systems. Effect of

More information

OVERHEADS ENHANCEMENT IN MUTIPLE PROCESSING SYSTEMS BY ANURAG REDDY GANKAT KARTHIK REDDY AKKATI

OVERHEADS ENHANCEMENT IN MUTIPLE PROCESSING SYSTEMS BY ANURAG REDDY GANKAT KARTHIK REDDY AKKATI CMPE 655- MULTIPLE PROCESSOR SYSTEMS OVERHEADS ENHANCEMENT IN MUTIPLE PROCESSING SYSTEMS BY ANURAG REDDY GANKAT KARTHIK REDDY AKKATI What is MULTI PROCESSING?? Multiprocessing is the coordinated processing

More information

A RESOURCE MANAGEMENT FRAMEWORK FOR INTERACTIVE GRIDS

A RESOURCE MANAGEMENT FRAMEWORK FOR INTERACTIVE GRIDS A RESOURCE MANAGEMENT FRAMEWORK FOR INTERACTIVE GRIDS Raj Kumar, Vanish Talwar, Sujoy Basu Hewlett-Packard Labs 1501 Page Mill Road, MS 1181 Palo Alto, CA 94304 USA { raj.kumar,vanish.talwar,sujoy.basu}@hp.com

More information

Complexity results for throughput and latency optimization of replicated and data-parallel workflows

Complexity results for throughput and latency optimization of replicated and data-parallel workflows Complexity results for throughput and latency optimization of replicated and data-parallel workflows Anne Benoit and Yves Robert GRAAL team, LIP École Normale Supérieure de Lyon June 2007 Anne.Benoit@ens-lyon.fr

More information

Parallelization Strategy

Parallelization Strategy COSC 6374 Parallel Computation Algorithm structure Spring 2008 Parallelization Strategy Finding Concurrency Structure the problem to expose exploitable concurrency Algorithm Structure Supporting Structure

More information

Introduction to parallel computers and parallel programming. Introduction to parallel computersand parallel programming p. 1

Introduction to parallel computers and parallel programming. Introduction to parallel computersand parallel programming p. 1 Introduction to parallel computers and parallel programming Introduction to parallel computersand parallel programming p. 1 Content A quick overview of morden parallel hardware Parallelism within a chip

More information

Lecture 9: Load Balancing & Resource Allocation

Lecture 9: Load Balancing & Resource Allocation Lecture 9: Load Balancing & Resource Allocation Introduction Moler s law, Sullivan s theorem give upper bounds on the speed-up that can be achieved using multiple processors. But to get these need to efficiently

More information

Boundary control : Access Controls: An access control mechanism processes users request for resources in three steps: Identification:

Boundary control : Access Controls: An access control mechanism processes users request for resources in three steps: Identification: Application control : Boundary control : Access Controls: These controls restrict use of computer system resources to authorized users, limit the actions authorized users can taker with these resources,

More information

Optimize HPC - Application Efficiency on Many Core Systems

Optimize HPC - Application Efficiency on Many Core Systems Meet the experts Optimize HPC - Application Efficiency on Many Core Systems 2018 Arm Limited Florent Lebeau 27 March 2018 2 2018 Arm Limited Speedup Multithreading and scalability I wrote my program to

More information

Chapter 3. Design of Grid Scheduler. 3.1 Introduction

Chapter 3. Design of Grid Scheduler. 3.1 Introduction Chapter 3 Design of Grid Scheduler The scheduler component of the grid is responsible to prepare the job ques for grid resources. The research in design of grid schedulers has given various topologies

More information

A Level-wise Priority Based Task Scheduling for Heterogeneous Systems

A Level-wise Priority Based Task Scheduling for Heterogeneous Systems International Journal of Information and Education Technology, Vol., No. 5, December A Level-wise Priority Based Task Scheduling for Heterogeneous Systems R. Eswari and S. Nickolas, Member IACSIT Abstract

More information

CSCE 626 Experimental Evaluation.

CSCE 626 Experimental Evaluation. CSCE 626 Experimental Evaluation http://parasol.tamu.edu Introduction This lecture discusses how to properly design an experimental setup, measure and analyze the performance of parallel algorithms you

More information

16/10/2008. Today s menu. PRAM Algorithms. What do you program? What do you expect from a model? Basic ideas for the machine model

16/10/2008. Today s menu. PRAM Algorithms. What do you program? What do you expect from a model? Basic ideas for the machine model Today s menu 1. What do you program? Parallel complexity and algorithms PRAM Algorithms 2. The PRAM Model Definition Metrics and notations Brent s principle A few simple algorithms & concepts» Parallel

More information

Data Model Considerations for Radar Systems

Data Model Considerations for Radar Systems WHITEPAPER Data Model Considerations for Radar Systems Executive Summary The market demands that today s radar systems be designed to keep up with a rapidly changing threat environment, adapt to new technologies,

More information

Non-uniform memory access machine or (NUMA) is a system where the memory access time to any region of memory is not the same for all processors.

Non-uniform memory access machine or (NUMA) is a system where the memory access time to any region of memory is not the same for all processors. CS 320 Ch. 17 Parallel Processing Multiple Processor Organization The author makes the statement: "Processors execute programs by executing machine instructions in a sequence one at a time." He also says

More information

IBM Spectrum Protect Version Introduction to Data Protection Solutions IBM

IBM Spectrum Protect Version Introduction to Data Protection Solutions IBM IBM Spectrum Protect Version 8.1.2 Introduction to Data Protection Solutions IBM IBM Spectrum Protect Version 8.1.2 Introduction to Data Protection Solutions IBM Note: Before you use this information

More information

A Data-Aware Resource Broker for Data Grids

A Data-Aware Resource Broker for Data Grids A Data-Aware Resource Broker for Data Grids Huy Le, Paul Coddington, and Andrew L. Wendelborn School of Computer Science, University of Adelaide Adelaide, SA 5005, Australia {paulc,andrew}@cs.adelaide.edu.au

More information

White Paper. Major Performance Tuning Considerations for Weblogic Server

White Paper. Major Performance Tuning Considerations for Weblogic Server White Paper Major Performance Tuning Considerations for Weblogic Server Table of Contents Introduction and Background Information... 2 Understanding the Performance Objectives... 3 Measuring your Performance

More information

Multimedia Systems 2011/2012

Multimedia Systems 2011/2012 Multimedia Systems 2011/2012 System Architecture Prof. Dr. Paul Müller University of Kaiserslautern Department of Computer Science Integrated Communication Systems ICSY http://www.icsy.de Sitemap 2 Hardware

More information

ECE 669 Parallel Computer Architecture

ECE 669 Parallel Computer Architecture ECE 669 Parallel Computer Architecture Lecture 9 Workload Evaluation Outline Evaluation of applications is important Simulation of sample data sets provides important information Working sets indicate

More information

IBM Tivoli Storage Manager Version Introduction to Data Protection Solutions IBM

IBM Tivoli Storage Manager Version Introduction to Data Protection Solutions IBM IBM Tivoli Storage Manager Version 7.1.6 Introduction to Data Protection Solutions IBM IBM Tivoli Storage Manager Version 7.1.6 Introduction to Data Protection Solutions IBM Note: Before you use this

More information

CSC630/CSC730 Parallel & Distributed Computing

CSC630/CSC730 Parallel & Distributed Computing CSC630/CSC730 Parallel & Distributed Computing Analytical Modeling of Parallel Programs Chapter 5 1 Contents Sources of Parallel Overhead Performance Metrics Granularity and Data Mapping Scalability 2

More information

Computer Architecture Lecture 27: Multiprocessors. Prof. Onur Mutlu Carnegie Mellon University Spring 2015, 4/6/2015

Computer Architecture Lecture 27: Multiprocessors. Prof. Onur Mutlu Carnegie Mellon University Spring 2015, 4/6/2015 18-447 Computer Architecture Lecture 27: Multiprocessors Prof. Onur Mutlu Carnegie Mellon University Spring 2015, 4/6/2015 Assignments Lab 7 out Due April 17 HW 6 Due Friday (April 10) Midterm II April

More information

Chapter 13 Strong Scaling

Chapter 13 Strong Scaling Chapter 13 Strong Scaling Part I. Preliminaries Part II. Tightly Coupled Multicore Chapter 6. Parallel Loops Chapter 7. Parallel Loop Schedules Chapter 8. Parallel Reduction Chapter 9. Reduction Variables

More information

Complexity Results for Throughput and Latency Optimization of Replicated and Data-parallel Workflows

Complexity Results for Throughput and Latency Optimization of Replicated and Data-parallel Workflows Complexity Results for Throughput and Latency Optimization of Replicated and Data-parallel Workflows Anne Benoit and Yves Robert GRAAL team, LIP École Normale Supérieure de Lyon September 2007 Anne.Benoit@ens-lyon.fr

More information

Analytical Modeling of Parallel Systems. To accompany the text ``Introduction to Parallel Computing'', Addison Wesley, 2003.

Analytical Modeling of Parallel Systems. To accompany the text ``Introduction to Parallel Computing'', Addison Wesley, 2003. Analytical Modeling of Parallel Systems To accompany the text ``Introduction to Parallel Computing'', Addison Wesley, 2003. Topic Overview Sources of Overhead in Parallel Programs Performance Metrics for

More information

HEP replica management

HEP replica management Primary actor Goal in context Scope Level Stakeholders and interests Precondition Minimal guarantees Success guarantees Trigger Technology and data variations Priority Releases Response time Frequency

More information

Layer-Based Scheduling Algorithms for Multiprocessor-Tasks with Precedence Constraints

Layer-Based Scheduling Algorithms for Multiprocessor-Tasks with Precedence Constraints Layer-Based Scheduling Algorithms for Multiprocessor-Tasks with Precedence Constraints Jörg Dümmler, Raphael Kunis, and Gudula Rünger Chemnitz University of Technology, Department of Computer Science,

More information

A Framework for Space and Time Efficient Scheduling of Parallelism

A Framework for Space and Time Efficient Scheduling of Parallelism A Framework for Space and Time Efficient Scheduling of Parallelism Girija J. Narlikar Guy E. Blelloch December 996 CMU-CS-96-97 School of Computer Science Carnegie Mellon University Pittsburgh, PA 523

More information

L3.4. Data Management Techniques. Frederic Desprez Benjamin Isnard Johan Montagnat

L3.4. Data Management Techniques. Frederic Desprez Benjamin Isnard Johan Montagnat Grid Workflow Efficient Enactment for Data Intensive Applications L3.4 Data Management Techniques Authors : Eddy Caron Frederic Desprez Benjamin Isnard Johan Montagnat Summary : This document presents

More information

Abstract of the Book

Abstract of the Book Book Keywords IEEE 802.16, IEEE 802.16m, mobile WiMAX, 4G, IMT-Advanced, 3GPP LTE, 3GPP LTE-Advanced, Broadband Wireless, Wireless Communications, Cellular Systems, Network Architecture Abstract of the

More information

Client Server & Distributed System. A Basic Introduction

Client Server & Distributed System. A Basic Introduction Client Server & Distributed System A Basic Introduction 1 Client Server Architecture A network architecture in which each computer or process on the network is either a client or a server. Source: http://webopedia.lycos.com

More information

On Cluster Resource Allocation for Multiple Parallel Task Graphs

On Cluster Resource Allocation for Multiple Parallel Task Graphs On Cluster Resource Allocation for Multiple Parallel Task Graphs Henri Casanova Frédéric Desprez Frédéric Suter University of Hawai i at Manoa INRIA - LIP - ENS Lyon IN2P3 Computing Center, CNRS / IN2P3

More information

Oracle Enterprise Manager. 1 Before You Install. System Monitoring Plug-in for Oracle Unified Directory User's Guide Release 1.0

Oracle Enterprise Manager. 1 Before You Install. System Monitoring Plug-in for Oracle Unified Directory User's Guide Release 1.0 Oracle Enterprise Manager System Monitoring Plug-in for Oracle Unified Directory User's Guide Release 1.0 E24476-01 October 2011 The System Monitoring Plug-In for Oracle Unified Directory extends Oracle

More information

IBM Best Practices Working With Multiple CCM Applications Draft

IBM Best Practices Working With Multiple CCM Applications Draft Best Practices Working With Multiple CCM Applications. This document collects best practices to work with Multiple CCM applications in large size enterprise deployment topologies. Please see Best Practices

More information

Understanding Parallelism and the Limitations of Parallel Computing

Understanding Parallelism and the Limitations of Parallel Computing Understanding Parallelism and the Limitations of Parallel omputing Understanding Parallelism: Sequential work After 16 time steps: 4 cars Scalability Laws 2 Understanding Parallelism: Parallel work After

More information

Embedded Systems Dr. Santanu Chaudhury Department of Electrical Engineering Indian Institution of Technology, Delhi

Embedded Systems Dr. Santanu Chaudhury Department of Electrical Engineering Indian Institution of Technology, Delhi Embedded Systems Dr. Santanu Chaudhury Department of Electrical Engineering Indian Institution of Technology, Delhi Lecture - 34 Compilers for Embedded Systems Today, we shall look at the compilers, which

More information

Best Practices. Deploying Optim Performance Manager in large scale environments. IBM Optim Performance Manager Extended Edition V4.1.0.

Best Practices. Deploying Optim Performance Manager in large scale environments. IBM Optim Performance Manager Extended Edition V4.1.0. IBM Optim Performance Manager Extended Edition V4.1.0.1 Best Practices Deploying Optim Performance Manager in large scale environments Ute Baumbach (bmb@de.ibm.com) Optim Performance Manager Development

More information

Kismet: Parallel Speedup Estimates for Serial Programs

Kismet: Parallel Speedup Estimates for Serial Programs Kismet: Parallel Speedup Estimates for Serial Programs Donghwan Jeon, Saturnino Garcia, Chris Louie, and Michael Bedford Taylor Computer Science and Engineering University of California, San Diego 1 Questions

More information

High Performance Computing in C and C++

High Performance Computing in C and C++ High Performance Computing in C and C++ Rita Borgo Computer Science Department, Swansea University Announcement No change in lecture schedule: Timetable remains the same: Monday 1 to 2 Glyndwr C Friday

More information

CSE398: Network Systems Design

CSE398: Network Systems Design CSE398: Network Systems Design Instructor: Dr. Liang Cheng Department of Computer Science and Engineering P.C. Rossin College of Engineering & Applied Science Lehigh University March 14, 2005 Outline Classification

More information

A Top Catching Scheme Consistency Controlling in Hybrid P2P Network

A Top Catching Scheme Consistency Controlling in Hybrid P2P Network A Top Catching Scheme Consistency Controlling in Hybrid P2P Network V. Asha*1, P Ramesh Babu*2 M.Tech (CSE) Student Department of CSE, Priyadarshini Institute of Technology & Science, Chintalapudi, Guntur(Dist),

More information

GrADSoft and its Application Manager: An Execution Mechanism for Grid Applications

GrADSoft and its Application Manager: An Execution Mechanism for Grid Applications GrADSoft and its Application Manager: An Execution Mechanism for Grid Applications Authors Ken Kennedy, Mark Mazina, John Mellor-Crummey, Rice University Ruth Aydt, Celso Mendes, UIUC Holly Dail, Otto

More information

Efficient, Scalable, and Provenance-Aware Management of Linked Data

Efficient, Scalable, and Provenance-Aware Management of Linked Data Efficient, Scalable, and Provenance-Aware Management of Linked Data Marcin Wylot 1 Motivation and objectives of the research The proliferation of heterogeneous Linked Data on the Web requires data management

More information

SolidFire and Pure Storage Architectural Comparison

SolidFire and Pure Storage Architectural Comparison The All-Flash Array Built for the Next Generation Data Center SolidFire and Pure Storage Architectural Comparison June 2014 This document includes general information about Pure Storage architecture as

More information

CLOUD-SCALE FILE SYSTEMS

CLOUD-SCALE FILE SYSTEMS Data Management in the Cloud CLOUD-SCALE FILE SYSTEMS 92 Google File System (GFS) Designing a file system for the Cloud design assumptions design choices Architecture GFS Master GFS Chunkservers GFS Clients

More information

Chapter 5: Analytical Modelling of Parallel Programs

Chapter 5: Analytical Modelling of Parallel Programs Chapter 5: Analytical Modelling of Parallel Programs Introduction to Parallel Computing, Second Edition By Ananth Grama, Anshul Gupta, George Karypis, Vipin Kumar Contents 1. Sources of Overhead in Parallel

More information

Graph Partitioning for Scalable Distributed Graph Computations

Graph Partitioning for Scalable Distributed Graph Computations Graph Partitioning for Scalable Distributed Graph Computations Aydın Buluç ABuluc@lbl.gov Kamesh Madduri madduri@cse.psu.edu 10 th DIMACS Implementation Challenge, Graph Partitioning and Graph Clustering

More information

Tag a Tiny Aggregation Service for Ad-Hoc Sensor Networks. Samuel Madden, Michael Franklin, Joseph Hellerstein,Wei Hong UC Berkeley Usinex OSDI 02

Tag a Tiny Aggregation Service for Ad-Hoc Sensor Networks. Samuel Madden, Michael Franklin, Joseph Hellerstein,Wei Hong UC Berkeley Usinex OSDI 02 Tag a Tiny Aggregation Service for Ad-Hoc Sensor Networks Samuel Madden, Michael Franklin, Joseph Hellerstein,Wei Hong UC Berkeley Usinex OSDI 02 Outline Introduction The Tiny AGgregation Approach Aggregate

More information

Designing Issues For Distributed Computing System: An Empirical View

Designing Issues For Distributed Computing System: An Empirical View ISSN: 2278 0211 (Online) Designing Issues For Distributed Computing System: An Empirical View Dr. S.K Gandhi, Research Guide Department of Computer Science & Engineering, AISECT University, Bhopal (M.P),

More information

Design of Parallel Algorithms. Models of Parallel Computation

Design of Parallel Algorithms. Models of Parallel Computation + Design of Parallel Algorithms Models of Parallel Computation + Chapter Overview: Algorithms and Concurrency n Introduction to Parallel Algorithms n Tasks and Decomposition n Processes and Mapping n Processes

More information

Parallel DBMS. Parallel Database Systems. PDBS vs Distributed DBS. Types of Parallelism. Goals and Metrics Speedup. Types of Parallelism

Parallel DBMS. Parallel Database Systems. PDBS vs Distributed DBS. Types of Parallelism. Goals and Metrics Speedup. Types of Parallelism Parallel DBMS Parallel Database Systems CS5225 Parallel DB 1 Uniprocessor technology has reached its limit Difficult to build machines powerful enough to meet the CPU and I/O demands of DBMS serving large

More information

Adaptive Aggregation Scheduling Using. Aggregation-degree Control in Sensor Network

Adaptive Aggregation Scheduling Using. Aggregation-degree Control in Sensor Network Contemporary Engineering Sciences, Vol. 7, 2014, no. 14, 725-730 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ces.2014.4681 Adaptive Aggregation Scheduling Using Aggregation-degree Control in

More information

1 Dulcian, Inc., 2001 All rights reserved. Oracle9i Data Warehouse Review. Agenda

1 Dulcian, Inc., 2001 All rights reserved. Oracle9i Data Warehouse Review. Agenda Agenda Oracle9i Warehouse Review Dulcian, Inc. Oracle9i Server OLAP Server Analytical SQL Mining ETL Infrastructure 9i Warehouse Builder Oracle 9i Server Overview E-Business Intelligence Platform 9i Server:

More information

Complexity Measures for Map-Reduce, and Comparison to Parallel Computing

Complexity Measures for Map-Reduce, and Comparison to Parallel Computing Complexity Measures for Map-Reduce, and Comparison to Parallel Computing Ashish Goel Stanford University and Twitter Kamesh Munagala Duke University and Twitter November 11, 2012 The programming paradigm

More information

SCHISM: A WORKLOAD-DRIVEN APPROACH TO DATABASE REPLICATION AND PARTITIONING

SCHISM: A WORKLOAD-DRIVEN APPROACH TO DATABASE REPLICATION AND PARTITIONING SCHISM: A WORKLOAD-DRIVEN APPROACH TO DATABASE REPLICATION AND PARTITIONING ZEYNEP KORKMAZ CS742 - PARALLEL AND DISTRIBUTED DATABASE SYSTEMS UNIVERSITY OF WATERLOO OUTLINE. Background 2. What is Schism?

More information

Analyzing the Performance of IWAVE on a Cluster using HPCToolkit

Analyzing the Performance of IWAVE on a Cluster using HPCToolkit Analyzing the Performance of IWAVE on a Cluster using HPCToolkit John Mellor-Crummey and Laksono Adhianto Department of Computer Science Rice University {johnmc,laksono}@rice.edu TRIP Meeting March 30,

More information

A Partial Critical Path Based Approach for Grid Workflow Scheduling

A Partial Critical Path Based Approach for Grid Workflow Scheduling A Partial Critical Path Based Approach for Grid Workflow Scheduling Anagha Sharaf 1, Suguna.M 2 PG Scholar, Department of IT, S.N.S College of Technology, Coimbatore, Tamilnadu, India 1 Associate Professor,

More information

Performance Extrapolation for Load Testing Results of Mixture of Applications

Performance Extrapolation for Load Testing Results of Mixture of Applications Performance Extrapolation for Load Testing Results of Mixture of Applications Subhasri Duttagupta, Manoj Nambiar Tata Innovation Labs, Performance Engineering Research Center Tata Consulting Services Mumbai,

More information

Grid Architectural Models

Grid Architectural Models Grid Architectural Models Computational Grids - A computational Grid aggregates the processing power from a distributed collection of systems - This type of Grid is primarily composed of low powered computers

More information

Grid Scheduler. Grid Information Service. Local Resource Manager L l Resource Manager. Single CPU (Time Shared Allocation) (Space Shared Allocation)

Grid Scheduler. Grid Information Service. Local Resource Manager L l Resource Manager. Single CPU (Time Shared Allocation) (Space Shared Allocation) Scheduling on the Grid 1 2 Grid Scheduling Architecture User Application Grid Scheduler Grid Information Service Local Resource Manager Local Resource Manager Local L l Resource Manager 2100 2100 2100

More information

An Introduction to Parallel Programming

An Introduction to Parallel Programming An Introduction to Parallel Programming Ing. Andrea Marongiu (a.marongiu@unibo.it) Includes slides from Multicore Programming Primer course at Massachusetts Institute of Technology (MIT) by Prof. SamanAmarasinghe

More information

TensorFlow: A System for Learning-Scale Machine Learning. Google Brain

TensorFlow: A System for Learning-Scale Machine Learning. Google Brain TensorFlow: A System for Learning-Scale Machine Learning Google Brain The Problem Machine learning is everywhere This is in large part due to: 1. Invention of more sophisticated machine learning models

More information

Lecture 7: Parallel Processing

Lecture 7: Parallel Processing Lecture 7: Parallel Processing Introduction and motivation Architecture classification Performance evaluation Interconnection network Zebo Peng, IDA, LiTH 1 Performance Improvement Reduction of instruction

More information

Mitigating Data Skew Using Map Reduce Application

Mitigating Data Skew Using Map Reduce Application Ms. Archana P.M Mitigating Data Skew Using Map Reduce Application Mr. Malathesh S.H 4 th sem, M.Tech (C.S.E) Associate Professor C.S.E Dept. M.S.E.C, V.T.U Bangalore, India archanaanil062@gmail.com M.S.E.C,

More information

Issues in Parallel Processing. Lecture for CPSC 5155 Edward Bosworth, Ph.D. Computer Science Department Columbus State University

Issues in Parallel Processing. Lecture for CPSC 5155 Edward Bosworth, Ph.D. Computer Science Department Columbus State University Issues in Parallel Processing Lecture for CPSC 5155 Edward Bosworth, Ph.D. Computer Science Department Columbus State University Introduction Goal: connecting multiple computers to get higher performance

More information

A Survey on Grid Scheduling Systems

A Survey on Grid Scheduling Systems Technical Report Report #: SJTU_CS_TR_200309001 A Survey on Grid Scheduling Systems Yanmin Zhu and Lionel M Ni Cite this paper: Yanmin Zhu, Lionel M. Ni, A Survey on Grid Scheduling Systems, Technical

More information

Enterprise print management in VMware Horizon

Enterprise print management in VMware Horizon Enterprise print management in VMware Horizon Introduction: Embracing and Extending VMware Horizon Tricerat Simplify Printing enhances the capabilities of VMware Horizon environments by enabling reliable

More information

Overview SENTINET 3.1

Overview SENTINET 3.1 Overview SENTINET 3.1 Overview 1 Contents Introduction... 2 Customer Benefits... 3 Development and Test... 3 Production and Operations... 4 Architecture... 5 Technology Stack... 7 Features Summary... 7

More information

Load Balancing for Parallel Multi-core Machines with Non-Uniform Communication Costs

Load Balancing for Parallel Multi-core Machines with Non-Uniform Communication Costs Load Balancing for Parallel Multi-core Machines with Non-Uniform Communication Costs Laércio Lima Pilla llpilla@inf.ufrgs.br LIG Laboratory INRIA Grenoble University Grenoble, France Institute of Informatics

More information

Microsoft SQL Server Fix Pack 15. Reference IBM

Microsoft SQL Server Fix Pack 15. Reference IBM Microsoft SQL Server 6.3.1 Fix Pack 15 Reference IBM Microsoft SQL Server 6.3.1 Fix Pack 15 Reference IBM Note Before using this information and the product it supports, read the information in Notices

More information

Scalable Computing: Practice and Experience Volume 10, Number 4, pp

Scalable Computing: Practice and Experience Volume 10, Number 4, pp Scalable Computing: Practice and Experience Volume 10, Number 4, pp. 413 418. http://www.scpe.org ISSN 1895-1767 c 2009 SCPE MULTI-APPLICATION BAG OF JOBS FOR INTERACTIVE AND ON-DEMAND COMPUTING BRANKO

More information

Temporal Exception Prediction for Loops in Resource Constrained Concurrent Workflows

Temporal Exception Prediction for Loops in Resource Constrained Concurrent Workflows emporal Exception Prediction for Loops in Resource Constrained Concurrent Workflows Iok-Fai Leong, Yain-Whar Si Faculty of Science and echnology, University of Macau {henryifl, fstasp}@umac.mo Abstract

More information

Compressing and Decoding Term Statistics Time Series

Compressing and Decoding Term Statistics Time Series Compressing and Decoding Term Statistics Time Series Jinfeng Rao 1,XingNiu 1,andJimmyLin 2(B) 1 University of Maryland, College Park, USA {jinfeng,xingniu}@cs.umd.edu 2 University of Waterloo, Waterloo,

More information

JAVA IEEE TRANSACTION ON CLOUD COMPUTING. 1. ITJCC01 Nebula: Distributed Edge Cloud for Data Intensive Computing

JAVA IEEE TRANSACTION ON CLOUD COMPUTING. 1. ITJCC01 Nebula: Distributed Edge Cloud for Data Intensive Computing JAVA IEEE TRANSACTION ON CLOUD COMPUTING 1. ITJCC01 Nebula: Distributed Edge for Data Intensive Computing 2. ITJCC02 A semi-automatic and trustworthy scheme for continuous cloud service certification 3.

More information

UNIT I (Two Marks Questions & Answers)

UNIT I (Two Marks Questions & Answers) UNIT I (Two Marks Questions & Answers) Discuss the different ways how instruction set architecture can be classified? Stack Architecture,Accumulator Architecture, Register-Memory Architecture,Register-

More information

DISTRIBUTED HIGH-SPEED COMPUTING OF MULTIMEDIA DATA

DISTRIBUTED HIGH-SPEED COMPUTING OF MULTIMEDIA DATA DISTRIBUTED HIGH-SPEED COMPUTING OF MULTIMEDIA DATA M. GAUS, G. R. JOUBERT, O. KAO, S. RIEDEL AND S. STAPEL Technical University of Clausthal, Department of Computer Science Julius-Albert-Str. 4, 38678

More information

Corral: A Glide-in Based Service for Resource Provisioning

Corral: A Glide-in Based Service for Resource Provisioning : A Glide-in Based Service for Resource Provisioning Gideon Juve USC Information Sciences Institute juve@usc.edu Outline Throughput Applications Grid Computing Multi-level scheduling and Glideins Example:

More information

Parallel Programming Concepts. Parallel Algorithms. Peter Tröger

Parallel Programming Concepts. Parallel Algorithms. Peter Tröger Parallel Programming Concepts Parallel Algorithms Peter Tröger Sources: Ian Foster. Designing and Building Parallel Programs. Addison-Wesley. 1995. Mattson, Timothy G.; S, Beverly A.; ers,; Massingill,

More information

A Taxonomy and Survey of Grid Resource Management Systems

A Taxonomy and Survey of Grid Resource Management Systems A Taxonomy and Survey of Grid Resource Management Systems Klaus Krauter 1, Rajkumar Buyya 2, and Muthucumaru Maheswaran 1 Advanced Networking Research Laboratory 1 Department of Computer Science University

More information

Mapping a group of jobs in the error recovery of the Grid-based workflow within SLA context

Mapping a group of jobs in the error recovery of the Grid-based workflow within SLA context Mapping a group of jobs in the error recovery of the Grid-based workflow within SLA context Dang Minh Quan International University in Germany School of Information Technology Bruchsal 76646, Germany quandm@upb.de

More information

Example: CPU-bound process that would run for 100 quanta continuously 1, 2, 4, 8, 16, 32, 64 (only 37 required for last run) Needs only 7 swaps

Example: CPU-bound process that would run for 100 quanta continuously 1, 2, 4, 8, 16, 32, 64 (only 37 required for last run) Needs only 7 swaps Interactive Scheduling Algorithms Continued o Priority Scheduling Introduction Round-robin assumes all processes are equal often not the case Assign a priority to each process, and always choose the process

More information

A Study of High Performance Computing and the Cray SV1 Supercomputer. Michael Sullivan TJHSST Class of 2004

A Study of High Performance Computing and the Cray SV1 Supercomputer. Michael Sullivan TJHSST Class of 2004 A Study of High Performance Computing and the Cray SV1 Supercomputer Michael Sullivan TJHSST Class of 2004 June 2004 0.1 Introduction A supercomputer is a device for turning compute-bound problems into

More information

Kenneth A. Hawick P. D. Coddington H. A. James

Kenneth A. Hawick P. D. Coddington H. A. James Student: Vidar Tulinius Email: vidarot@brandeis.edu Distributed frameworks and parallel algorithms for processing large-scale geographic data Kenneth A. Hawick P. D. Coddington H. A. James Overview Increasing

More information

Introduction to Distributed Systems. INF5040/9040 Autumn 2018 Lecturer: Eli Gjørven (ifi/uio)

Introduction to Distributed Systems. INF5040/9040 Autumn 2018 Lecturer: Eli Gjørven (ifi/uio) Introduction to Distributed Systems INF5040/9040 Autumn 2018 Lecturer: Eli Gjørven (ifi/uio) August 28, 2018 Outline Definition of a distributed system Goals of a distributed system Implications of distributed

More information

CS6030 Cloud Computing. Acknowledgements. Today s Topics. Intro to Cloud Computing 10/20/15. Ajay Gupta, WMU-CS. WiSe Lab

CS6030 Cloud Computing. Acknowledgements. Today s Topics. Intro to Cloud Computing 10/20/15. Ajay Gupta, WMU-CS. WiSe Lab CS6030 Cloud Computing Ajay Gupta B239, CEAS Computer Science Department Western Michigan University ajay.gupta@wmich.edu 276-3104 1 Acknowledgements I have liberally borrowed these slides and material

More information

Solution of Out-of-Core Lower-Upper Decomposition for Complex Valued Matrices

Solution of Out-of-Core Lower-Upper Decomposition for Complex Valued Matrices Solution of Out-of-Core Lower-Upper Decomposition for Complex Valued Matrices Marianne Spurrier and Joe Swartz, Lockheed Martin Corp. and ruce lack, Cray Inc. ASTRACT: Matrix decomposition and solution

More information

A Scalable Content- Addressable Network

A Scalable Content- Addressable Network A Scalable Content- Addressable Network In Proceedings of ACM SIGCOMM 2001 S. Ratnasamy, P. Francis, M. Handley, R. Karp, S. Shenker Presented by L.G. Alex Sung 9th March 2005 for CS856 1 Outline CAN basics

More information

THE GLOBUS PROJECT. White Paper. GridFTP. Universal Data Transfer for the Grid

THE GLOBUS PROJECT. White Paper. GridFTP. Universal Data Transfer for the Grid THE GLOBUS PROJECT White Paper GridFTP Universal Data Transfer for the Grid WHITE PAPER GridFTP Universal Data Transfer for the Grid September 5, 2000 Copyright 2000, The University of Chicago and The

More information

Memory Management Algorithms on Distributed Systems. Katie Becker and David Rodgers CS425 April 15, 2005

Memory Management Algorithms on Distributed Systems. Katie Becker and David Rodgers CS425 April 15, 2005 Memory Management Algorithms on Distributed Systems Katie Becker and David Rodgers CS425 April 15, 2005 Table of Contents 1. Introduction 2. Coarse Grained Memory 2.1. Bottlenecks 2.2. Simulations 2.3.

More information

Verification and Validation of X-Sim: A Trace-Based Simulator

Verification and Validation of X-Sim: A Trace-Based Simulator http://www.cse.wustl.edu/~jain/cse567-06/ftp/xsim/index.html 1 of 11 Verification and Validation of X-Sim: A Trace-Based Simulator Saurabh Gayen, sg3@wustl.edu Abstract X-Sim is a trace-based simulator

More information

Auto Management for Apache Kafka and Distributed Stateful System in General

Auto Management for Apache Kafka and Distributed Stateful System in General Auto Management for Apache Kafka and Distributed Stateful System in General Jiangjie (Becket) Qin Data Infrastructure @LinkedIn GIAC 2017, 12/23/17@Shanghai Agenda Kafka introduction and terminologies

More information

Improving the Efficiency of Fast Using Semantic Similarity Algorithm

Improving the Efficiency of Fast Using Semantic Similarity Algorithm International Journal of Scientific and Research Publications, Volume 4, Issue 1, January 2014 1 Improving the Efficiency of Fast Using Semantic Similarity Algorithm D.KARTHIKA 1, S. DIVAKAR 2 Final year

More information

Anna Morajko.

Anna Morajko. Performance analysis and tuning of parallel/distributed applications Anna Morajko Anna.Morajko@uab.es 26 05 2008 Introduction Main research projects Develop techniques and tools for application performance

More information

Principles of Parallel Algorithm Design: Concurrency and Mapping

Principles of Parallel Algorithm Design: Concurrency and Mapping Principles of Parallel Algorithm Design: Concurrency and Mapping John Mellor-Crummey Department of Computer Science Rice University johnmc@rice.edu COMP 422/534 Lecture 3 17 January 2017 Last Thursday

More information