Replication and scheduling Methods Based on Prediction in Data Grid

Similar documents
Dynamic Data Grid Replication Strategy Based on Internet Hierarchy

Nowadays data-intensive applications play a

A Novel Data Replication Policy in Data Grid

MANAGEMENT AND PLACEMENT OF REPLICAS IN A HIERARCHICAL DATA GRID

Future Generation Computer Systems. PDDRA: A new pre-fetching based dynamic data replication algorithm in data grids

RDT: A New Data Replication Algorithm for Hierarchical Data Grid

TAXONOMY OF OPTIMIZATION APPROACHES OF RESOURCE BROKERS IN DATA GRIDS

A Dynamic Replication Strategy based on Exponential Growth/Decay Rate

EDPFRS: ENHANCED DYNAMIC POPULAR FILE REPLICATION AND SCHEDULING FOR DATA GRID ENVIRONMENT

Improving Data Grids Performance by using Modified Dynamic Hierarchical Replication Strategy

A new efficient Virtual Machine load balancing Algorithm for a cloud computing environment

Simulation model and instrument to evaluate replication technologies

NEW REPLICA SELECTION TECHNIQUE FOR BINDING CHEAPEST REPLICA SITES IN DATA GRIDS

Future Generation Computer Systems. A survey of dynamic replication strategies for improving data availability in data grids

A Survey on Resource Allocation policies in Mobile ad-hoc Computational Network

A Set Coverage-based Mapping Heuristic for Scheduling Distributed Data-Intensive Applications on Global Grids

RPLB: A Replica Placement Algorithm in Data Grid with Load Balancing

Simulation of Dynamic Data Replication Strategies in Data Grids

IJCSI International Journal of Computer Science Issues, Vol. 9, Issue 5, No 2, September 2012 ISSN (Online):

DYNAMIC DATA STORAGE AND PLACEMENT SYSTEM BASED ON THE CATEGORY AND POPULARITY

Available online at ScienceDirect. Procedia Computer Science 78 (2016 ) A Dynamic Data Replication in Grid System

Automatic Data Placement and Replication in Grids

Addressed Issue. P2P What are we looking at? What is Peer-to-Peer? What can databases do for P2P? What can databases do for P2P?

Mapping a group of jobs in the error recovery of the Grid-based workflow within SLA context

CS473: Course Review CS-473. Luo Si Department of Computer Science Purdue University

Department of Physics & Astronomy

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

Tortoise vs. hare: a case for slow and steady retrieval of large files

Improving the Performance of the Peer to Peer Network by Introducing an Assortment of Methods

Evaluation of an Economy-Based File Replication Strategy for a Data Grid

Adaptive replica consistency policy for Kafka

Data location-aware job scheduling in the grid. Application to the GridWay metascheduler

Online Optimization of VM Deployment in IaaS Cloud

OVERHEADS ENHANCEMENT IN MUTIPLE PROCESSING SYSTEMS BY ANURAG REDDY GANKAT KARTHIK REDDY AKKATI

World Academy of Science, Engineering and Technology International Journal of Computer and Information Engineering Vol:2, No:4, 2008

Introductions. Computer Networking Lecture 01. January 16, HKU SPACE Community College. HKU SPACE CC CN Lecture 01 1/36

CS 421: COMPUTER NETWORKS SPRING FINAL May 21, minutes

A Novel Replication Strategy in Data Grid Environment with a Dynamic Threshold

A Heuristic Algorithm for Designing Logical Topologies in Packet Networks with Wavelength Routing

Link Lifetime Prediction in Mobile Ad-Hoc Network Using Curve Fitting Method

A Time-To-Live Based Reservation Algorithm on Fully Decentralized Resource Discovery in Grid Computing

Experience of Data Grid simulation packages using.

Performance Analysis of Data Mining Classification Techniques

Survey on MapReduce Scheduling Algorithms

An Experimental Cloud Resource Broker System for Virtual Application Control with VM Allocation Scheme

Load Balancing Algorithm over a Distributed Cloud Network

Research Article Combining Pre-fetching and Intelligent Caching Technique (SVM) to Predict Attractive Tourist Places

Performing MapReduce on Data Centers with Hierarchical Structures

A ROUTING MECHANISM BASED ON SOCIAL NETWORKS AND BETWEENNESS CENTRALITY IN DELAY-TOLERANT NETWORKS

A Self-Adaptive Insert Strategy for Content-Based Multidimensional Database Storage

Mining Distributed Frequent Itemset with Hadoop

A New Intelligent Method in Brokers to Improve Resource Recovery Methods in Grid Computing Network

Co-operative Scheduled Energy Aware Load-Balancing technique for an Efficient Computational Cloud

CS 425 / ECE 428 Distributed Systems Fall 2015

A Replication and Cache based Distributed Metadata Management System for Data Grid

Rank Measures for Ordering

QoS Guided Min-Mean Task Scheduling Algorithm for Scheduling Dr.G.K.Kamalam

Low Overhead Geometric On-demand Routing Protocol for Mobile Ad Hoc Networks

Connectivity, Energy and Mobility Driven Clustering Algorithm for Mobile Ad Hoc Networks

A General Data Grid: Framework and Implementation

Improving the Efficiency of Fast Using Semantic Similarity Algorithm

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

A Resource Discovery Algorithm in Mobile Grid Computing Based on IP-Paging Scheme

Analysis of Cluster based Routing Algorithms in Wireless Sensor Networks using NS2 simulator

A New HadoopBased Network Management System with Policy Approach

ADAPTIVE AND DYNAMIC LOAD BALANCING METHODOLOGIES FOR DISTRIBUTED ENVIRONMENT

Simulation of a cost model response requests for replication in data grid environment

Distributed Two-way Trees for File Replication on Demand

Multimedia Streaming. Mike Zink

Archna Rani [1], Dr. Manu Pratap Singh [2] Research Scholar [1], Dr. B.R. Ambedkar University, Agra [2] India

Scaling Distributed Machine Learning with the Parameter Server

Correlative Analytic Methods in Large Scale Network Infrastructure Hariharan Krishnaswamy Senior Principal Engineer Dell EMC

Grid Computing Systems: A Survey and Taxonomy

Localized and Incremental Monitoring of Reverse Nearest Neighbor Queries in Wireless Sensor Networks 1

A Data-Aware Resource Broker for Data Grids

Correlation based File Prefetching Approach for Hadoop

Introduction: Two motivating examples for the analytical approach

Resource CoAllocation for Scheduling Tasks with Dependencies, in Grid

Dynamic Data Placement Strategy in MapReduce-styled Data Processing Platform Hua-Ci WANG 1,a,*, Cai CHEN 2,b,*, Yi LIANG 3,c

Web Data mining-a Research area in Web usage mining

Information Retrieval and Web Search Engines

Study of Load Balancing Schemes over a Video on Demand System

INTRODUCTION TO ICT.

Predicting Scientific Grid Data Transfer Characteristics

Performance Analysis of Applying Replica Selection Technology for Data Grid Environments*

Chapter The LRU* WWW proxy cache document replacement algorithm

Overlapping Ring Monitoring Algorithm in TIPC

Intelligent management of on-line video learning resources supported by Web-mining technology based on the practical application of VOD

Partial Caching Scheme for Streaming Multimedia Data in Ad-hoc Network

Centralized versus distributed schedulers for multiple bag-of-task applications

A Comparative Study of Selected Classification Algorithms of Data Mining

CLOUD-SCALE FILE SYSTEMS

International Journal of Advanced Research in Computer Science and Software Engineering

Chapter 3. Design of Grid Scheduler. 3.1 Introduction

Job Re-Packing for Enhancing the Performance of Gang Scheduling

A Set Coverage-based Mapping Heuristic for Scheduling Distributed Data-Intensive Applications on Global Grids

New Optimal Load Allocation for Scheduling Divisible Data Grid Applications

IJREAT International Journal of Research in Engineering & Advanced Technology, Volume 1, Issue 5, Oct-Nov, ISSN:

Routing Protocols in MANETs

Lecture on Computer Networks

Transcription:

Australian Journal of Basic and Applied Sciences, 5(11): 1485-1496, 2011 ISSN 1991-8178 Replication and scheduling Methods Based on Prediction in Data Grid 1 R. Sepahvand, 2 A. Horri and 3 Gh. Dastghaibyfard 1-3 Department of Computer Science and Engineering, College of Engineering, Shiraz University, Molla Sadra Ave, Shiraz, Iran. 71348-51154. Abstract: Nowadays, data intensive applications such as HEP (High Energy Physics), Genetics, Nuclear Physics and are common. These applications generate a huge amount of data. The size of data in such applications is TB or PB and these data must be shared among researchers all around the world. Data grid can help to manage this amount of data. Data grid provides services for the sharing and managing of large data files around the world. One of the major data grid goals is to reduce job response time. To achieve this objective, the main challenge is network bandwidth. One common method to tackle network traffic is to replicate files at different sites and then, to schedule the jobs to the sites which are close to the files required for the job execution. In order to replicate files, it is required that best replica be selected to reduce access latency. On the other hand, because of network traffic dynamism, computing the accurate file transfer time is time-consuming or even impossible; hence other methods such as prediction can be used to estimate the file transfer time. In this paper, a cosine similarity predictor function is proposed that uses the path similarity between the previous and current file transfer paths. Based on this predictor function, new replication and scheduling algorithms are proposed. Experimental results obtain by Optorsim show that the proposed replication and scheduling algorithms outperform multi-regression and neural network methods as the number of jobs increases. Key words: Replication, Prediction, scheduling, Optorsim, Data Grid. INTRODUCTION Data grid is a type of grid, used in the data-intensive applications where the size of data files reaches TB or sometimes PB. High Energy Physics (HEP), Genetics and Earth Observation are such examples. In such applications, managing and storing this huge data is very difficult or even impossible. Data grid tries to distribute data files in several sites. Therefore, most applications have the choice to retrieve data files, from various sites. The main factors influencing the performance of data grid are the latency of the network (especially in WAN environments) and the bandwidth between the sites. Prior to execution, a data grid job commonly requests many files that may be available at various sites. In order to facilitate file transfers in data grid environment, the names of the files as well as the corresponding sites that save these files are kept in a replica catalog. Here, there are three questions to be addressed: 1. The first question arises as to how the best site should be selected to obtain replicas? Since the requested files may be available at various sites, the best site is the one with the minimum file transfer time. For each path from the available sites (sources) to the current node (destination), the accurate file transfer time should be computed. In order to compute the accurate file transfer time for each path, it is required that the network bandwidth of all nodes participating from the sources to the destination be obtained. This is a timeconsuming or even impossible task, leading to the use of the estimated time. In our approach, each site records its recent file transfer times and paths (last k minutes, k = 10, 20, 30...) in the log and uses this log to estimate the file transfer time. 2. The second question has to do with what file deletion strategy should be applied. When there is lack of storage in a site for replica files, numerous methods are available for file deletion, namely LRU, LFU and the like. 3. The third question is to do with how the e appropriate site should be selected to execute the jobs? To schedule jobs, static attributes (the type of operating system, LAN and WAN network bandwidth, processor speed, and storage capacity and secondary memory) and dynamic attributes (processor utilization, physical memory utilization, free secondary memory size, current usage price, and network bandwidth utilization) must be considered. Then the appropriate site must be selected based on the requestor criteria. The rest of the paper is organized as follows. Section 2 presents the related work. Section 3 provides the motivation as well as the proposed replication and deletion algorithms. Experimental results are given in Section 4, and finally, Section 5 concerns the paper conclusion. Corresponding Author: A. Horri, Department of Computer Science and Engineering, Colege of Engineering, Shiraz University, Molla Sadra Ave, Shiraz, Iran. 71348-51154. E-mail: dstghaib@shirazu.ac.ir 1485

Related Work: In (A Novel Job Scheduling Algorithm for Improving Data Grid's Performance) we proposed a novel scheduling algorithm and replication strategy. In this algorithm, the decision as to where to execute a job is made by considering the job requirements, length of job queue, and the current grid status. In (K. Sashi, A.S. Thanamani, 2011) a Modified BHR Algorithm is proposed. This Algorithm replicates the region files in to the region header and stories them in the site where the file has frequently been accessed. Mohamed et al., (2004) have proposed Close-to-Files (CF) algorithm that schedules a job to least loaded processors close to a site where data are present. Assuming that a job only needs a single file as input data, CF algorithm uses an exhaustive search across all combinations of computing sites and data sites to find a combination with the minimum cost, including computation and transmission delay. Data replication is also used to improve performance. They claim simulation results show that CF can achieve good performance when compared to Worst-Fit job placement algorithm, which places jobs on the execution sites with the largest number of idle processors. Chakrabarti et al., (2004) have proposed Integrated Replication and Scheduling Strategy (IRS) that decouples jobs scheduling from data scheduling. At the end of periodic intervals the popularity of files is calculated and then used by the data scheduler to replicate data for the next set of jobs, which may or may not share the same data requirements as the previous set. Dang et al., (2007) in their proposed algorithm organize the data in to several data categories. Then, this category is used for improving data replication placement strategy. For replicating a file they maintain data accesses locally and globally and compute average number of access. For each file, if it is accessed more than the average access, it will be replicated. We (R. Sepahvand et al., 2008) have proposed a replication and scheduling algorithm for a 3-level hierarchical structure. For replication a list of candidate sites that hold the file is proposed, and then a site with the highest bandwidth to the requested node is selected. Should the available space in SE (Storage Element) be greater or equal to the requested file size, the file will be replicated. Past data transfer time of files in each site that hold replicas can provide a reasonable approximation of the end-to-end transfer throughput. Deelman et al., (2002) have proposed a replication algorithm that uses a cost model to predict whether replicas are worth creating. It is found to be more effective in reducing average job time than the basic case where there is no replication. The test bed for their simulation is a combination of ring and fat-tree structure where leaf nodes are CE (computing elements) and other nodes are SE (storage elemnts). Li et al., (2007) have proposed a new replica selection algorithm that uses markov chain to achieve state transition probability matrix to predict the reliability of replicas in the form of probability by the system state classification and Analyzing the factors affecting replica selection. The grey system theory is utilized to help predicting the data response time on the basis the GM (1,1) grey dynamic model; They showed that replica selection model, the prediction is valid, helpful to selection decision, and is able to achieve load balance between nodes holding replicas. Vazkhudai and Schopf (2001;2002;2003) have proposed multivariate regression techniques that uses past data transfer time for better predictions. By using data from data streams of network, file size, and past grid transfer information, they achieve a better accuracy in file transfer throughput prediction. Rahman et al., (2008) have proposed an algorithm that uses file history and two predictor functions namely multi regression and neural network. To select the best replica they use k-nearest Neighbor (KNN) rule. The KNN rule uses past file transfer logs. They showed neural network (with back propagation, to train the network,) predictive technique outperforms multi regression model. Kim et al., (2008) have presented a decentralized, resource selection techniques based on accessibility. Their techniques rely on local, historic observations. They claim the estimation techniques are sufficiently accurate to provide a meaningful rank order of nodes based on their accessibility. All the predicting file transfer time algorithms in the literature require pervious transfer time from source to destination (i.e. exact match); otherwise they are unable to predict file transfer time especially if suddenly a link fails due to dynamic nature of grid. When in file transfer history, exact match for replica is not available, or exact match is availabe but some links have failed, then partial match is a good prediction method that is explained in the next section. Motivation and Proposed Algorithms: As explained in previous section, use of partial match is an appropriate method for predecting file transfer time. More log data can lead to a better prediction based on current network traffic. It is shown in the next section that the last k minutes log for k=30 is the best case. Since ftp (a connected oriented) protocol is used for file transfer, in this paper file transfer path is modeled as a vector, where each element of this vector represent a node in the grid. for determining similarity between two vectors, cosine similarity method is used that is explained next. 1486

Cosine Similarity: One way to measure similarity between two vectors with n dimensions is to compute their cosine angle i.e. cosine similarity. Formula 1, shows how to compute cosine similarity CS (A, B) of two vectors A (x1, x2,., xn) and B (y1, y2,, yn), and it is depicted in figure. A B CS(A,B) A B x1* x2 y1* y2 2 2 2 2 ( x1 y1 ) * ( x2 y2 ) Fig. 1: Cosine similarity between two vectors. The resulting similarity ranges from -1 to 1, where -1 means exactly opposite and 1 means exactly the same. In-between values indicate intermediate dissimilarity or similarity. Proposed File Transfer Time Predictor: When different sites hold replicas of a particular file, the access latency can be minimized by selecting the best replica. In this paper, a predictive technique is proposed to estimate the file transfer time for different sites that hold replicas of a file. The site that has the lowest estimated transfer time is selected to fetch replica.ion III. To measure the similarity of two paths, two factors are considered; the common sub-routes between the paths and the bandwidth of each part of the network. Each path is represented as a vector V, whose length is equal to the number of all existing edges in the network graph. To compute the cosine similarity for two paths x, y CS ( x, y ), Suppose that network has n edges, each edge has a number between 0 n-1, also the length of each vector is n and each edge has an element in vector. e i j The value of each edge, (edge from node i to node j) in vector V, i.e. V[ e i, j ] is equal to inverse of s s bandwidth from site i to site j or 0 if the related edge does not exist in the path. The inverse value is used because file transfer time is more sensitive to low bandwidth. In other words, the presence of a common low-bandwidth edge in two paths leads to a higher similarity between their transfer times. (the buildvector function in predictor pseudo code works like this.) When site s s i transfer a file from site j, path and time of this transfer is saved in the local catalog of s site i. s The following pseudo code is used to compute predictor function for site i. Each site Si keeps an LPT array, where LPT[j] is the predicted transfer time from Sj to Si that calculated by weighting average time of three path in Si s catalog that have maximum similarity with current path. The weight of each path that is used for weighting is its similarity with current path. maxi is ith maximum CS (v,vi) for all vi in recent (last k minutes) local catalog of site Si. Predictor () { // compute estimated transfer time from s // all sites to site i foreach site s j (j i) { s s vector V = buildvector ( i, j ) 1487

LPT[j] = average ( max 1, max 2, max 3 ) //maxi is ith maximum CS (v,vi) for all vi in recent // (last k minutes) local catalog of site Si. } } // end Predictor This method can be used as general traffic parameters estimator such as round trip time (RTT) in network. The Proposed Algorithms: In this section a dynamic method for deleting and replication file and job scheduling are proposed. For each site s i, let: LPT[j] is the time to transfer a file with constant size (ConstSize) from site sj to site si. for computing LPT[j] we use path transfer time exist in catalog and cosine similarity method. The detail has been show in predictor function code in section B. TFi,j is the time to transfer a file from site si to site sj, TFi,j can be computed as follow: TFi,j = LPT[j]*Filesize ConstSize Proposed File Deletion Algorithm: The best case in data grid is when local storage has infinite space for replicating all files. But this not always possible, therefore deleting file is a must. Different algorithms have been proposed for deleting files, like LRU and LFU. A file that has replica in other sites with less predicted transfer time is more appropriate for deletion. Since if later we need such a file, it can be transferred immediately. Also a file that has not been used recently probably will not be used in near feature too. So such a file is more appropriate for deletion. In other words, the proposed algorithm deletes a file that has a near replica and has not been used recently. For each file i in a site, a delete factor (df i ) is computed as follow: k 2 df i = k 1 ru i + mtf i Where: k 1 and k 2 are the weights for each factor, which represent their importance. (k 1 +k 2 = 1) ru i is recently used file access time mtf i is minimum predicted transfer time for file i from other sites and stored in a dft (delete factor) table. A file with highest df i will be selected for deletion. Proposed Job Scheduling Algorithm: For scheduling two important parameters should be considered: estimated file transfer time and queue length. Data grid jobs have high file transfer time so reducing file transfer time will reduce overall completion time. Therefore selecting a site with minimum transfer time is a must. Queue length, i.e. the number of jobs that are waiting in queue affects the overall completion time. To select a proper site, a parallel scheduling algorithm is proposed as follow. s i has minimum estimated file transferring time for a job but has a high queue length and another site s j is idle with more estimated files transferring time for this job, in this case we should estimate transferring time for remaining files (files needed for jobs that are in Queue and haven t transfer yet) and declare this time plus estimated transferring time for new job to scheduler. Since in data grid, transferring time is more than computing time and this little time is overlapped with transferring time, we ignore computing time. The following pseudo code is used to scheduling algorithm: Master: // Master broadcasts job to all CE's: S BCAST ( i JK) // receive from each CE (Slave), sum of files transfer //time estimated with predictor function 1488

S Receive ( i, T i ) // Find appropriate site for execution the job i.e. the //site that has least time ( T ) Site Sj = Site with min i // Send message to site Sj for executing the job Send (Sj, Run) Salve: // Receive a job name from master Receive (Master, JK) F //Get logical file names m Fn From job //description T // Now compute i = estimated sum of transfer time //of files F m Fn to current site and the file transfe //time for jobs in the queue with predictor function //that use cosine similarity and with: T = i n f m min( TF i, j ) //TFij=Time to transfer Logical file f from site j to site i T Send (Master, i ) The execution of a master and slave sites for the cosine similarity prediction are illustrated in figure.2 and figure.3. Fig. 2: Execution flow of master. Fig. 3: Execution flow of slave. 1489

Proposed Replication Algorithm: Assume job P has been assigned to site s i for execution and requires K input files. If the minimum transfer time is below a certain time-threshold, the file will be accessed remotely (S.M. Park et al., 2003). // pseudo replication algorithm For file f = 1 to K do { Min = If file f is not available in site s i { For each site s j that has a copy of file f { If ( TF, < Min) { i j Min = TF i, j S = j } } // end for each if (Min < time-threshold) then Accesses file f remotely from site S Else If (enough space exist in site s i ) then Replicate file f from site S Else { Apply delete algorithm; Replicate file f from site S } // end else } // end if } // end if } // end for Each of the proposed algorithms is dynamic. Dynamic means in every k minutes intervals the estimated file transfer time for a constant file size (ConstSize) will be updated in LPT (Local Predicted Table). Fig. 4: Replication flow. 1490

Experimental Results: To evaluate the proposed algorithms, Optorsim (Optorsim- A Replica Optimiser Simulation; D.G. Cameron et al.,; W.H. Bell et al., 2003) is used as the simulation tool and the EU data grid as network, which are explained next. Optorsim: The components of Optorsim are as follow and also depicted in Figure 5. Resource Scheduler: It receives jobs from user and sends them to the best node according to proposed algorithm. Storage Element (SE): Storage resource in grid. Computing Element (CE): Computing resource in grid. Replica Manager: It controls data transferring in each node and provide a mechanism for accessing the Data Catalog. Replica Optimizer: It has replica algorithm and control file replication according to proposed algorithm. Based on the scheduling algorithm the broker sends jobs to a node. Each job needs a list of files to run. Reducing file access time is the final objective of optimization algorithms. EU Data Grid: EU Data Grid Network (The DataGrid Project) (See Figure 6) consists of a number of sites. Each site has zero or more SE (Storage Element) and zero or more CE (Computing Element). CE's provide computational resource and SE's provide storage elements. Brokers will assign jobs to sites that have CE's. A job usually requires a set of files which some may be available on the assigned CE. So those files that are not available should be located by replica manager first and then decide to access them remotely or replicate them on the assigned CE. Fig. 5: Simulator architecture (S.M. Park et al., 2003). Fig. 6: The EU Data Grid and their associated network (R.M. Rahmana et al., 2008). 1491

Site 0 in Figure 6, is the CERN (European Organization for Nuclear Research) location and contains original copy of some data sample files that cannot be deleted. Since all files are available in Site 0, so any job assigned to this site does not need any file transfer. Therefore in our simulation we only consider all CE sites except site 0. In Figure 6 sites without CE are shown as star and have the following properties: Have high storage capacity. Do not provide any computing power to outside. Act as intermediate nodes to replicate files for faster access. Do not generate any workload. The bandwidth between sites is expressed in M and G (Mbits/Sec or Gbits/Sec respectively). The input for simulation is in three configuration files. File one contains network graph (nodes and edges), for each node, number and size of each SE, number of CE's, and for each edge its bandwidth. File two contains a file table (filename, size and index), job table (job name and required files), CE schedule table (CE id and the jobs that can execute on that CE) and job selection probability. File three contains simulation parameters such as: number of jobs to be run, scheduling and optimization algorithm, file access pattern, In this simulation we used the same configurations that is used in (R.M. Rahmana et al., 2008) and is common in Optorsim, i.e. six types of jobs with no overlapping between the set of files each job accesses. Each file is assumed to be 1 GB. To record file transfer time and path, we modified Optorsim code. A job will typically request a set of logical filename (s) for data access. The order in which the files are requested is determined by the access pattern. In this simulation, we experiment with five access patterns that are shown in Figure 7: Sequential: files are accessed in the order that has been stated in the job configuration file. Random: files are accessed using flat random distribution. Unitary random: file requests are one element away from previous file requests but the direction will be random. Gaussian random walk: files are accessed using a Gaussian distribution. Zipf distribution: most of the requests in the web follow Zipf distribution. Fig. 7: Various file access patterns (S.M. Park et al., 2003). 1492

Simulation Results: To evaluate the predictor function, the best way is comparing the time that the function has estimated for the file transfer (before transferring the file) with actual file transfer time (after transferring the file). To evaluate the prediction method, the following steps were used: 1. Modifying the Optorsim code, for the recording of path and transfer time between different sites. 2. Classifying the result of step 1 based on the destination site. 3. Developing a software for the implementing of our cosine similarity prediction method. 4. Selecting 10% of each data site predicting the transfer time of these data with 90% of the remaining data, and repeating this for the remaining data. 5. comparing the predicted time with the actual time. In our simulation, for the steps 4 and 5 we used various time intervals from 15 to 60 minutes and noticed that the decreasing of the interval time will lead to insufficient information about the path for prediction and the increasing of the interval time will lead to the bad estimation of the current network traffic. In particular results for 30 minutes time the interval was the best. In the predictor function constant file size of 10 GB was used to compute the transfer time. In deleting files, we use the combination of predicting and LRU. LRU is used for estimating future use and predicting is used for deleting files that have replica in near site. This two factors cause to reduce total completion time. Fig. 6 depicts the % or correct predictions by the cosine similarity for different access patterns. Here The % of correct predictions is calculated by the prediction time divided by the actual time. Gaussian 1 0.98 0.96 0.94 0.92 0.9 0.88 S2 S4 S5 S6 S7 S8 S11 S13 S15 S16 Random 1 0.995 0.99 0.985 0.98 0.975 0.97 0.965 0.96 S2 S4 S5 S6 S7 S8 S11 S13 S15 S16 1 Zipf 0.99 0.98 0.97 0.96 0.95 0.94 S2 S4 S5 S6 S7 S8 S11 S13 S15 S16 1493

1 0.995 0.99 0.985 0.98 0.975 0.97 0.965 0.96 0.955 Sequential S2 S4 S5 S6 S7 S8 S11 S13 S15 S16 Unitary 1 0.99 0.98 0.97 0.96 0.95 0.94 0.93 S2 S4 S5 S6 S7 S8 S11 S13 S15 S16 Fig. 8: %of correct prediction for different access pattern. Figure 8 compares the three prediction methods i.e. multi-variant regression, neural network and cosine similarity. It is clear that cosine similarity outperforms the other two methods. If the available storage for replication is not enough, our proposed algorithm will only replicate those files that are not available in near sites (with transfer time more than threshold) and delete files that have less predicted transfer time, so it will not delete those files that have a high cost of transfer. Therefore this method has better performance comparing to LRU. But if the available storage for replication is enough, or there is a high bandwidth between source and destination, both algorithms have the same performance. 1 0.8 0.6 0.4 0.2 0 Sequential Gaussian Random Zipf Unitary REG NN Cosine Similarity Fig. 9: Comparison cosine similarity method with regression and Neural Network methods. Method Scheduling Replication QL (Red) Queue length LRU QR (Blue) Queue length Proposed replication SR (Green) Proposed scheduling Proposed replication Fig. 10: Compares following combinations for scheduling and replication. In QL and QR methods, even though same scheduling is used, QR has a better performance over 20% due to proposed replication algorithm. In QR and SR methods, even though same replication is used, SR has a better performance over 20% due to proposed scheduling algorithm. 1494

When the number of job increases the performance of our method also increases comparing with other method. In grid environment a lot of job should be run. Fig. 10: Mean run time. Conclusion: In this paper, a new predictor function is proposed to estimate the file transfer time, which uses the path similarity between the previous and current file transfer paths. Based on this predictor function, three algorithms are proposed for replication, deletion and job scheduling. We analyzed the performance of the cosine similarity replication method as well as the regression and neural network model for five file access patterns. Experimental results with Optorsim showed that our replica algorithm outperforms the multi-regression and neural network methods. Also, the deletion algorithm has a better performance as compared to the traditional LRU deletion method as the number of jobs increases. REFERENCES A Novel Job Scheduling Algorithm for Improving Data Grid's Performance to Sixth International Conference on P2P, Parallel, Grid, Cloud and Internet Computing. Bell, W.H., D.G. Cameron, L. Capozza, A.P. Millar, K. Stockinger and F. Zini, 2003. OptorSim- a Grid simulator for studying dynamic data replication strategies, International Journal of High Performance Computing Applications, 17(4): 403-416. Cameron, D.G., A.P. Millar, C. Nicholson, R. Carvajal-Schiaffino, F. Zini and K. Stockinger. A SIMULATION TOOL FOR SCHEDULING AND REPLICA OPTIMISATION IN DATA GRIDS, www.gridpp.ac.uk/papers/chep04_optorsim. Chakrabarti, A., R.A. Dheepak and S. Engupta, 2004. Integration of scheduling and replication in Data Grids, in: Lecture Notes in Computer Science, 3296: 375-385. Dang, N.N. and S.B. Lim, 2007. Combination of Replication and Scheduling in Data Grids, IJCSNS International Journal of Computer Science and Network Security, 7 (3): 304-308. Deelman, E., H. Lamehamedi, B. Szymanski and S. Zujun, 2002. Data replication strategies in grid environments, in: Proceedings of 5th International Conference on Algorithms and Architecture for Parallel Processing, ICA3PP 2002, IEEE Computer Science Press, Bejing, China, pp: 378-383. Jinoh Kim, Abhishek Chandra and Jon B. Weissman, 2008. Accessibility-based Resource Selection in Loosely-coupled Distributed Systems Distributed Computing Systems, 2008. Li, J., 2007. A Replica Selection Approach based on Prediction in Data Grid, Third International Conference on Semantics, Knowledge and Grid (SKG 2007),Xian, China, (s): 274-277. Mohamed, H.H. and D.H.J. Epema, 2004. An evaluation of the close-to-files processor and data coallocation policy in multiclusters, in: 2004 IEEE International Conference on Cluster Computing, IEEE Society Press, San Diego, California, USA, pp: 287-298. Optorsim - A Replica Optimiser Simulation, http://grid-data-management.web.cern.ch/grid-datamanagement/optimisation/optor/ Park, S.M., J.H. Kim and Y.B. Ko, 2003. Dynamic Grid Replication Strategy based on Internet Hierarchy, Proceedings of the second International Workshop on Grid and cooperative Computing (GCC2003), Shanghai, China. Rahmana, R.M., R. Alhajja and K. Barkera, 2008. Replica selection strategies in data grid, Journal of Parallel and Distributed Computing, 68(12): 1561-1574. 1495

Sashi, K. and A.S. Thanamani, 2011. Dynamic replication in a data grid using a Modified BHR region based algorithm, Future Generation Computer Systems, 27: 202-210. Sepahvand, R. and A. Horri and Gh. Dastghaibyfard, 2008. A new 3-layer replication and scheduling strategy In data grid, International Symposium on Telecommunications (IST 2008), Tehran, Iran, pp: 27-28. The DataGrid Project. http://www.edu-datagrid.org Vazhkudai, S. and J. Schopf, 2003. Using regression techniques to predict large data Transfers, Computing: Infrastructure and Applications, The International Journal of High Performance Computing Applications, IJHPCA, 17(3): 249-268. Vazhkudai, S., J. Schopf and I. Foster, 2002. Predicting the Performance of Wide-Area Data Transfers, 16 th International Parallel and Distributed Processing Symposium (IPDPS 2002), Fort Lauderdale, FL, USA. Vazhkudai, S., S. Tuecke and I. Foster, 2001. Replica selection in the globus data grid, First IEEE/ACM International Conference on Cluster Computing and the Grid (CCGrid 2001), Brisbane, Australia, 2001. 1496