A Low-Cost Multi-Failure Resilient Replication Scheme for High Data Availability in Cloud Storage

Size: px
Start display at page:

Download "A Low-Cost Multi-Failure Resilient Replication Scheme for High Data Availability in Cloud Storage"

Transcription

1 216 IEEE 23rd International Conference on High Perforance Coputing A Low-Cost Multi-Failure Resilient Replication Schee for High Data Availability in Cloud Storage Jinwei Liu* and Haiying Shen *Departent of Electrical and Coputer Engineering, Cleson University, Cleson, SC 29634, USA Departent of Coputer Science, University of Virginia, Charlottesville, VA 2294, USA jinweil@cleson.edu, hs6s@virginia.edu Abstract Replication is a coon approach to enhance data availability in cloud storage systes. Previously proposed replication schees cannot effectively handle both correlated and noncorrelated achine failures while increasing the data availability with the liited resource. The schees for correlated achine failures ust create a constant nuber of replicas for each data object, which neglects diverse data popularities and cannot utilize the resource to axiize the expected data availability. Also, the previous schees neglect the consistency aintenance cost and the storage cost caused by replication. It is critical for cloud providers to axiize data availability (hence iniize SLA violations) while iniizing cost caused by replication in order to axiize the revenue. In this paper, we build a nonlinear integer prograing odel to axiize data availability in both types of failures and iniize the cost caused by replication. Based on the odel s solution for the replication degree of each data object, we propose a low-cost ulti-failure resilient replication schee (). can effectively handle both correlated and non-correlated achine failures, considers data popularities to enhance data availability, and also tries to iniize consistency aintenance cost and storage cost. Extensive nuerical results fro trace paraeters and experients fro real-world Aazon S3 show that achieves high data availability, low data loss probability and low consistency aintenance cost and storage cost copared to previous replication schees. I. INTRODUCTION Datacenter storage syste (e.g., Hadoop Distributed File Syste (HDFS) [1], RAMCloud [2], Google File Syste (GFS) [3] and Windows Azure [4]) is an iportant coponent of cloud datacenters, especially for data-intensive services in this big data era. It is critical for cloud providers to reduce the violations of Service Level Agreeents (SLAs) for tenants to provide high quality-of-service and avoid the associated penalties. For exaple, a typical SLA required fro services that use Aazon s Dynao storage syste is that 99.9% of the read and write requests execute within 3s [5]. Therefore, a storage syste ust guarantee data availability for different applications to coply to SLAs, which is, however, a non-trivial task. Iany cloud services such as Aazon, Google App Engine and Windows Azure, the server failure probability is in the range of [1%, 1%] and the corresponding data availability is in the range of [99.9%, 99.99%] [6]. Data availability is usually influenced by data loss, which is typically caused by achine failures including correlated and non-correlated achine failures coexisting in storage systes. The forer eans ultiple nodes fail (nearly) siultaneously, while the latter eans nodes fail individually. Correlated achine failures often occur in large-scale storage systes [7] [9] due to coon failure causes (e.g., cluster power outages, workload-triggered software bug anifestations, Denial-of- Service attacks). For exaple, in cluster power outages [1], [11], a non-negligible percentage (.5%-1%) of nodes [1], [1] do not coe back to life after power is restored [12]. Correlated achine failures cause significant data loss [12], which have been docuented by Yahoo! [1], LinkedIn [1] and Facebook [13]. Non-correlated achine failures [8] are caused by reasons such as different hardware/software copositions and configurations, and varying network access ability. Replication is a coon approach to reduce data loss and enhance data availability. The data popularity consideration in replication is critical to axiize the expected data availability 1. Due to highly skewed data popularity distributions [16], popular data with considerably higher request frequency generates heavier load on the nodes, which ay lead to data unavailability at a tie. On the other hand, unpopular data with few requests wastes the resources for storing and aintaining replicas. Thus, an effective replication schee ust consider the diverse data popularities to use the liited resources (e.g., eory) [17], [18] to increase expected data availability. Though creating ore replicas for data objects iproves data availability, it coes with higher consistency aintenance cost [19] [21] and storage cost [22], [23]. In consistency aintenance, when data is updated, an update is sent to its replica nodes. In addition to the nuber of replicas, the consistency aintenance cost and storage cost are also affected by the geographic distance and storage edia (e.g., disk, SSD, EBS), respectively. To reduce the cost, we need to iniize the nuber of replicas, liit the geographic distance of replica nodes and replicate the additional replicas beyond the required storage of applications to less expensive storage edia while still provide SLA guarantee. Therefore, effective replication ethods ust axiize expected data availability (by considering both correlated and non-correlated achine failures and data popularity) and iniize the cost caused by replication (i.e., consistency aintenance cost and storage cost) [2], [24]. Rando replication has been widely used in datacenter storage systes including HDFS, RAMCloud, GFS and Windows Azure. These systes partition each data object into chunks (i.e., partitions), each of which is replicated to a constant nuber of randoly selected nodes on different racks. Though ran- 1 We use a service-centric availability etric: the proportion of all successful service requests over all requests [8], [14], [15] /16 $ IEEE DOI 1.119/HiPC

2 do replication can handle non-correlated achine failures, it cannot handle correlated achine failures well. The reason is that any cobination of a certain nuber of nodes would for a set of nodes for storing all replicas of a data chunk and data loss occurs if all nodes in the set experience failures siultaneously [12]. Previously proposed data replication ethods cannot effectively handle both correlated and noncorrelated achine failures and utilize the liited resource to increase data availability siultaneously [7], [8], [12], [25]. Although any ethods have been proposed to iprove data availability [7], [8], [19], [25] [27], they do not concurrently consider data popularities and the cost caused by replication to increase expected data availability and decrease cost. As shown in Figure 1, a key proble here is how to achieve an optial tradeoff between increasing data availability and reducing cost caused by replication with the ultiate goal of SLA copliance and revenue Data availability Copyset Replication Rando Replication Consistency aintenance cost Figure 1: Achieving an optial tradeoff aong data availability, storage cost and consistency aintenance cost. axiization for cloud service providers. To address the proble, we propose a low-cost ulti-failure resilient replication schee (), which targets the application of distributed data-intensive services and storage syste. is ore advantageous than previous replication schees (e.g., Rando Replication, Copyset Replication [12]) in that it can handle both correlated and non-correlated achine failures, and also jointly considers data popularity and the cost caused by replication. We suarize the contribution of this work below. We build a nonlinear integer prograing (NLIP) odel that ais to axiize expected data availability (in both types of failures) with the consideration of popularities and reduce the cost caused by replication. We then use Lagrange ultipliers to derive a solution: the replication degree (i.e., the nuber of replicas) of each data object. Based on the solution, we propose to handle both correlated and non-correlated achine failures. partitions all nodes to different groups with each group responsible for replicating all data objects with the sae replication degree. Then, partitions nodes in a group into different sets [12]; each set consists of nodes fro different datacenters within a certain geographic distance. The replicas of a chunk are stored in the nodes in one set, so data loss occurs only if these nodes experience failures siultaneously. Each data chunk is replicated to the corresponding storage ediu based on its priority. reduces the frequency of data loss by reducing the nuber of sets in a group, i.e., the probability that all nodes in a set fail. We have conducted extensive nuerical analysis based on trace paraeters and experients on Aazon S3 [28] to copare with other state-of-the-art replication schees. Results show achieves high data availability, low data loss probability and low storage and consistency aintenance cost. The reainder of this paper is organized as follows. Section II reviews the related work. Section III presents our NLIP odel. Section IV presents the design for. Section V presents the nuerical and experiental results. Section VI concludes this paper with rearks on our future work. II. RELATED WORK Many ethods have been proposed to handle non-correlated or correlated achine failures. Zhong et al. [8] assued independent achines failures, and proposed a odel that achieves high expected service availability. However, this odel does not consider correlated achine failures hence cannot handle such failures. Nath et al. [7] identified a set of design principles that syste builders can use to tolerate correlated failures. Cidon et al. proposed Copyset Replication [12] and Tiered Replication [29] to reduce the frequency of data loss caused by correlated achine failures by liiting the replica nodes of any chunks to a single copyset. These ethods for correlated achine failures do not consider data popularity to iniize data loss probability in correlated and non-correlated achine failures. Unlike Copyset Replication and Tiered Replication, our proposed reduces probability of data loss caused by both correlated and non-correlated achine failures with considering data popularity, and it derives diverse replication degrees for data objects with different popularities and thus increases the overall data object availability by creating ore replicas for popular data objects and less for unpopular data objects. Moreover, replicates data objects with considering reducing consistence aintenance cost and storage cost, which is critical for cloud providers to axiize the revenue. Thoson et al. [25] presented CalvinFS, a scalable distributed file syste, which horizontally partitions and replicates file syste etadata across a shared-nothing cluster of servers. However, CalvinFS cannot handle both correlated and non-correlated achine failures while iniizing cost (e.g., consistency aintenance cost and storage cost) caused by replication. The above previously proposed ethods cannot handle both correlated achine failures and non-correlated achine failures while utilizing the liited resource to increase data availability. They neglect data popularity existing in current cloud storage syste and thus cannot fully utilize the resource to increase data availability. Also, they do not concurrently consider iniizing the consistency aintenance cost and storage cost caused by replication. There is a large body of work on enhancing data availability for distributed storage systes. Abu-Libdeh et al. [26] presented a replication protocol for datacenter services to provide strong consistency and high availability. Bonvin et al. [19] proposed a self-anaged key-value store that dynaically allocates the resources of a data cloud to several applications in a cost-efficient and fair anner. Zhang et al. [3] presented Moji to provide the reliability and availability in large-scale storage systes. Moji uses a two-tier architecture in which the priary tier contains a irrored pair of nodes and the secondary tier contains the secondary backup nodes with weakly consistent copies of data. Yu et al. [31] used a fixed nuber of replicas for every 243

3 data object, and showed that the assignent of object replicas to achines plays a draatic role in the availability of ultiobject operations. However, ost of these previous works neglect data object popularities when deterining the nuber of replicas for each object, thus cannot satisfy the deands of popular data objects or fully utilize the liited resource. Although the works in [8], [16] consider object popularities, they do not give a solution for handling correlated achine failures, which is a key issue in achieving high availability in today s cloud datacenters [7]. In addition, they do not consider different storage ediu prices (for reducing storage cost) or geographic distance in selecting nodes to store replicas, both of which affect the total cost of the storage systes. Motivated by the probles in the existing works, we propose that can effectively handle both correlated and noncorrelated achine failures and also considers the factors indicated in the introduction section to axiize data availability and reduce the consistency aintenance cost and storage cost. III. NONLINEAR INTEGER PROGRAMMING MODEL FOR A cloud storage syste usually serves ultiple applications siultaneously. Without loss of generality, we assue there are n applications in the cloud storage, and each application has data objects [19]. Each data object belongs to only one application, and is split into M partitions [19] and the data object is lost if any of its partitions is lost [12]. Replication degree of a data object represents the nuber of replicas of the data object. We use D ij to denote the jth data object belonging to application i (denoted by a i ). Let d ij be the replication degree (nuber of replicas) of D ij. The replicas of a partition of a data object are placed in a set of d ij different nodes. Suppose there are N servers in the cloud. For analytical tractability, we assue a physical node (i.e., a server) belongs to a rack, a roo, a datacenter, a country and a continent. To easily identify the geographic location of a physical node, each physical node has a label in the for of continent-country-datacenter-roo-rack-server [19]. Proble Stateent: Given data object request probabilities, data object sizes, space constraints for different applications, and thresholds for request failure probability, consistency aintenance cost and storage cost, what is the optial replication degree for each data object, so that the request failure probability, consistency aintenance cost and storage cost are iniized in both correlated and noncorrelated achine failures? Then, how to assign the chunk replicas of data objects to the nodes to achieve the objectives? A. Data Availability Maxiization We axiize data availability by considering both achine failure probability and data object request probability (i.e., popularity). We iniize the data loss probability in both correlated and non-correlated achine failures. Different data objects ay have different popularities, and then have different deands on the nuber of replicas to ensure data availability. 1) Correlated Machine Failures: To reduce data loss in correlated achine failures, we adopt the fault-tolerant set (FTS) concept fro [12], which is a distinct set of servers holding all copies of a given partition. Each FTS is a single unit of failure. That is, when an FTS fails, at least one data object is lost. For correlated achine failures, the probability of data loss increases as the nuber of FTSs increases because the probability that the failed servers constitute at least one FTS increases (It is ore likely that the failed servers include at least one FTSs). So we iniize the probability of data loss by iniizing the nuber of FTSs. To this end, we can statically assign each server to a single FTS, and constrain the replicas of a partition to a randoly selected preassigned FTS. That is, we first place the priary replica (i.e., original copy) on a randoly selected server, and then place the secondary replicas on all the nodes in this server s FTS. However, this will lead to load ibalance proble in which soe servers becoe overloaded while soe servers are underloaded due to data storage. On the other hand, rando replication that randoly chooses a replica holder has a higher probability of data loss because alost every new replicated partition creates a distinct FTS. To achieve a tradeoff, we use the approach in Copyset Replication [12], which assigns a server to a liited nuber of FTSs rather than a single FTS. Due to the overlap of FTSs, after the priary replica is stored in a randoly selected server, the FTS options that can store the secondary replicas are those that hold the server. The servers in these sets are options that can be used to store the secondary replicas. The nuber of these servers is called scatter width (denoted by S). For exaple, if the FTSs that hold server 1 are {1,2,3}, {1,4,5}, then when the priary replica is stored at server 1, the secondary replica can be randoly placed either on servers 2 and 3 or 4 and 5. In this case the scatter width equals to 4. The probability of correlated achine failures equals the ratio of the nuber of FTSs over the axiu nuber of sets: #FTSs/ax{#sets} (1) To reduce the probability of data loss, Copyset Replication akes the replication schee satisfy two conditions below: Condition 1: The FTSs overlap with each other by at ost one server. Condition 2: The FTSs cover all the servers equally. Copyset Replication uses the Balanced Incoplete Block Design (BIBD)-based ethod for any scatter width to create FTSs that satisfy both condition 1 and 2 and iniize #FTSs. 2 We define a pair (X,A), where X is a set of servers in the syste (i.e., X = {1,2,...,N}), and A is a collection of all FTSs in the syste. Let N,R and λ be positive integers such that N > R 2, BIBD for (N,R,λ) satisfies the following properties: Each FTS contains exactly R servers. Each pair of servers is contained in exactly λ FTSs. When λ = 1, the BIBD provides an optial design for iniizing the nuber of FTSs for scatter width S = N 1. In this case, condition 1 ensures that each FTS increases the scatter width for its servers by exactly R 1 copared to the case S when λ =. Copyset Replication creates R 1 N R FTSs. Then, the failure probability in correlated achine failures equals: 2 In ipleentation, Copyset Replication uses rando perutation to create FTSs. 244

4 p cor = S ( ) N N R 1 R / (2) R In rando replication, the nuber of FTSs created is [12]: { ( #FTSs = N S R 1), S < N 2 #FTSs ( N) (3) R, S N Based on Equs. (1), (3), we can calculate the probability of correlated achine failures in rando replication. We give an exaple to illustrate the process of generating FTSs. Consider a storage syste with N = 12 servers, the size of FTS R = 3, and the scatter width S = 4. Using the BIBDbased ethod, the following solution of FTSs is created to achieve less nuber of FTSs. B 1 = {,1,2},B 2 = {3,4,1},B 3 = {6,7,8},B 4 = {9,1,11}, B 5 = {,3,8},B 6 = {1,4,7},B 7 = {2,5,11},B 8 = {5,6,9}. For rando replication, the #FTSs is 72. Hence, the probability of data loss caused ( ) by correlated ( ) achine failures is: N 12 #FTSs/ = 72/ =.327 R 3 However, for BIBD-based ethod fro Copyset Replication, the nuber of FTSs is 8, and the probability of data loss caused by correlated achine failures is uch saller: ( ) ( ) N 12 #FTSs/ = 8/ =.36 R 3 There are any ethods that can be used for constructing BIBDs, but no single ethod can create optial BIBDs for any cobination of N and R [32], [33]. Copyset Replication cobines BIBD and rando perutations to generate a nonoptial design that can accoodate any scatter width. By relaxing the constraint for the nuber of overlapping nodes fro an exact nuber to at ost the exact nuber, the probability of successfully generating FTSs increases. Since the scatter width should be uch saller than the nuber of nodes, is likely to generate FTSs with at ost one overlapping node [12]. tries to iniize the replication degrees of data objects in order to liit the consistency aintenance cost and storage cost, the replication degrees should be not large or vary greatly. Saller replication degrees generate saller R values. Although soeties it is not possible to generate the optial BIBDs, BIBDs can be used as a useful benchark to easure how close is to the optial schee for specific values of scatter width [12]. 2) Non-correlated Machine Failures: In non-correlated achine failures, the failures of achines are statistically independent of each other, and they can be categorized into unifor and nonunifor achine failures. In the scenario of unifor achine failures, each achine has the sae probability to fail, denoted by p ( < p < 1). The data object is lost if any of its chunk (partition) is lost, and a chunk is lost only if all replicas of the chunk are lost. Hence, the expected probability of data loss in the unifor achine failure is: p uni =( M p d ij )/( n) (4) where M is the nuber of partitions (chunks) for each data object, n is the nuber of applications, and is the nuber of data objects in each application. In the scenario of nonunifor achine failures, each achine fails with different probabilities. Assue replicas of each data object are placed oachines with no concern for individual achine failures. Let p 1,...,p N be the failure probabilities of N servers in the cloud, respectively. Based on [8], the expected data object failure probability is the sae as that on unifor failure achines with per-achine failure probability equaling N i=1 p i/n. We use p non to denote the expected probability of data loss in nonunifor achine failures. Then, N p non =( M ( p k /N) d ij )/( n) (5) k=1 We introduce a ethod to evaluate a data object s popularity below. Different types of applications are always featured by the popular tie periods of data objects. For exaple, the videos in social network applications are usually popular for 3 onths [34], while the news in a news website usually is popular for several days. We rank the applications based on their types and use b ai to denote the rank; a higher b ai eans that the application has longer popular tie periods of data objects. Assue tie is split into epochs (a fixed period of tie). To avoid creating and deleting replicas for data objects that are popular for a short tie period, we consider both its application rank (b ai ) and its expected visit frequency, i.e., the nuber of visits in an epoch (v ij ) [16], [19], [35], [36], to deterine the popularity function of a data object (ϕ( )): ϕ ij ( )=α b ai + β v ij (6) where α and β are the weights for the two factors. Let r ij be the probability of requesting data D ij. For singleobject requests, n i=1 r ij = 1 after noralization. The request probability of a data object is the sae as that of each partition of this data object. The request probability is proportional to the popularity of the data object (ϕ( )) [8], [37]. r ij = k 1 ϕ ij ( ) (7) where k 1 is a certain coefficient. 3) Availability Maxiization Proble Forulation: In a cloud syste with the co-existence of correlated and noncorrelated (unifor and nonunifor) achine failures, the data is lost if any type of failures occurs. Recall that p cor (Forula (2)), p uni (Forula (4)) and p non (Forula (5)) are probabilities of a data object loss caused by correlated achine failures, unifor achine failures, and nonunifor achine failures (at epoch t), respectively. So the probability of data loss caused by correlated and non-correlated achine failures is 3 P f = w 1 p cor + w 2 p uni + w 3 p non ( w i = 1) (8) i=1 where w 1,w 2, and w 3 are the probabilities of the occurrence of each type of achine failures, respectively. We axiize the data availability by iniizing the expected request failure probability with the consideration that different data objects have different popularities. To achieve this goal, we present a NLIP odel with ultiple constraints to deterine the replication degree of each data object. Given the popularities and sizes of data objects, our goal is to find the optial replication degree of each data object such that the expected request failure probability (denoted by P r ) can be iniized. Each application purchases a certain storage space in public clouds. Also, in private clouds, different applications ay be assigned different storage spaces based on their priorities. Thus, the storage space constraint for each application is necessary and iportant due 245

5 to the liited precious storage edia. Thus, we foralize our proble as the following constrained optiization proble: in P r = r ij M (P f ) d ij (9) s.t. s ij d ij K i (i = 1,...,n) where K i (i = 1,...,n) is the space capacity for application i, and s ij is the size of data object D ij. s ij d ij is the total storage consuption of all the data objects belonging to application i. The optiization objective is to iniize the expected request failure probability and the constraint is to ensure that the storage consuption of an application does not exceed its space capacity. B. Consistency Maintenance Cost Miniization In this section, we forulate the proble of iniizing consistency aintenance cost caused by data replication. We use C c to denote the total consistency aintenance cost of all replicas in the cloud storage syste. We first introduce how to calculate C c. Previous work [38] indicates that the data object write overhead is linear with the nuber of data object replicas. Then, the consistency aintenance cost of a partition can be approxiated as the product of the nuber of replicas of the partition and the average counication cost paraeter (denoted by δ co ) [19], i.e., d ij δ co. Hence, the total consistency aintenance cost of all data objects is C c = (M d ij ) δ co (1) The fixed average counication cost (δ co ) can be calculated as in [39]: δ co = E[ s u dis(s i,s j ) σ] (11) i, j where s u is the average update essage size, dis(s i,s j ) is the geographic distance between the server of the original copy S i and a replica server S j (an expectation of all possible distances between server of original copy (priary server) and replica servers, which is calculated it fro a probabilistic perspective), and σ is the average counication cost per unit of distance. We adopt the ethod in [19] to calculate the geographic distance between servers. Specifically, it uses a 6-bit nuber to represent the distance between servers. Each bit corresponds to the location part of a server, i.e., continent, country, datacenter, roo, rack and server. Starting with the ost significant bit (i.e., the leftost bit), each location part of both servers are copared one by one to copute the geo-siilarity between the. The corresponding bit is set to 1 if the location parts are equivalent, otherwise it is set to. Once a bit is set to, all less significant bits are autoatically set to. For exaple, suppose S i and S j are two arbitrary servers, and the distance between the is represented as 111 (as shown below). Then, it indicates that S i and S j are in the sae datacenter but not in the sae roo. continent country datacenter roo rack server The geographic distance is obtained by applying a binary NOT operation to the geo-siilarity. That is, 111 = 111 = 7 (decial) Thus, the optiization proble for consistency aintenance cost can be forulated as follows: in C c = (M d ij ) δ co s.t. s ij d ij K i (i = 1,...,n) (12) C. Storage Cost Miniization Different applications require different storage edia (e.g., disk, SSD, EBS) to store data. Different edia have different costs per unit size. For exaple, in the Aazon web cloud service, the prices for EBS and SSD are $.44 and $.7 per hour, respectively. After satisfying the applications storage requireents on different storage edia, we need to decide the storage edia for their additional replicas for enhanced data availability. Different applications have different SLAs with different associated penalties. The applications with higher SLA violation penalties should have higher priorities to eet their SLA requireents in order to reduce the associated penalties. Therefore, the additional replicas of data objects of higher-priority applications should be stored in a faster storage ediu. On the other hand, in order to save storage cost, the additional replicas of data objects of lower-priority applications should be stored in a lower and less expensive storage ediu. We use b pi to denote application i s priority; higher b pi eans higher priority. Since faster storage edius are ore costly, for data objects of high priority applications, we hope to store ore data partitions in faster storage edius in order to satisfy ore requests per unit tie hence increase data availability [19]. Thus, to deterine the storage ediu for storing additional replicas, we define a priority function ψ( ) of a data object based on its application priority and size [16]: ψ ij ( )=γ b pi + η/s ij (13) where γ and η are the weights for application priority and data object s size. The size of the data object can be changed due to write operation. The priority values are classified to a nuber of levels, and each level corresponds to a storage ediu. Thus, the unit storage cost of a data object equals the unit cost of its apped storage ediu, denoted by c sij, which is proportional to the priority value of a data object (ψ ij ( )). c sij = k 2 ψ ij ( ) (14) where k 2 is a certain coefficient. The storage cost of a data object is related to its storage ediu, its size and the nuber of its replicas. Different storage edius have different unit costs, and different data objects have different sizes and replication degrees. We iniize storage cost by iniizing the expected cost for storage edius. We exaine expected storage cost iniization by deterining the replication degree for each data object. To achieve this goal, we present a NLIP approach with ultiple constraints that can be used to obtain a policy. Given the applications of data objects and the data object sizes, our goal is to find the replication degree for each data object that iniizes storage cost. Hence, we foralize our proble as the following constrained optiization proble: in C s = (s ij d ij c sij T ) s.t. s ij d ij K i (i = 1,...,n) where T is the tie duration of an epoch. (15) 246

6 D. Proble Forulation and Solution We consider additional three threshold constraints for request failure probability (Pr th ), consistency aintenance cost (Cc th ) and storage cost (Cs th ). The probability of expected request failure ust be no larger than a threshold Pr th. This constraint is used to ensure that the expected request failure probability is not beyond a threshold, which serves the goal of achieving high data availability. The constraint on the consistency aintenance cost is to ensure that the consistency aintenance cost in one epoch is no larger than a threshold, Cc th. The storage cost in one epoch is no ore than threshold Cs th. The constraints on both consistency aintenance cost and storage cost are to ensure that the cost caused by replication is at a low level, which akes the syste ore efficient and econoical. By cobining Forulas (9), (12) and (15) and the additional constraints, we can build a NLIP odel for the global optiization proble as follows: in { P r +C c +C s } = (r ij M (P f ) d ij ) (16) + (M d ij ) δ co + (s ij d ij c sij T ) s.t. s ij d ij K i (i = 1,...,n) (17) r ij M (P f ) d ij Pr th ( < r ij < 1) (18) (M d ij ) δ co Cc th (19) (s ij d ij c sij T ) Cs th (2) The decision variables are data objects s replication degrees, and they ust be positive integers in practice. The objective is to iniize the request failure probability, the consistency aintenance cost and storage cost. The optiization constraints are used to ensure that the space capacity of data objects belonging to each application is not exceeded, the probability of expected request failure is no ore than a threshold P th r, the consistency aintenance cost and storage cost in one epoch are no ore than thresholds Cc th and Cs th, respectively. Theore 3.1. The relaxed NLIP optiizatioodel is convex. Proof Inequs. (17), (19), (2) are linear inequalities, and they define convex regions. The exponential function P d ij f in Inequ. (18) is convex, and the su (i.e., n i=1 r ij (P f ) d ij) of convex functions is a convex function. Thus, the constraint (18) defines a convex set. All the constraints define convex regions, and the intersection of convex sets is a convex set. Thus, the region of the optiization proble is convex. Hence the relaxed NLIP optiization proble is convex. For analytical tractability, we first relax the proble to a real-nuber optiization proble in which d 11,...,d n are real nubers, and derive the solution for the real-nuber optiization proble. Then, we use integer rounding to get the solution for practical use. Specifically, we adopt the approach fro [8] to round each d ij to its nearest integer while all d ij s saller than 1 are rounded to 1. By relaxing the proble to real-nuber optiization proble, the optial solution should always use up all the available storage space (i.e., K i ) for each application [8]. 3 Thus, we have s ij d ij = K i (i = 1,...,n) (21) where K i is the space capacity for application i. For the realnuber optiization proble, we use Lagrange ultipliers to derive the solution. Since there are n + 3 constraints, we use the ultipliers λ 1,λ 2,...,λ n+3 to cobine the constraints and the optiization goal together into the Lagrangian function Λ(d 11,...,d 1,...,d n1,...,d n,λ 1,λ 2,...,λ n+3 ) = ( r ij M (P f ) d ij + (M d ij ) δ co + (s ij d ij c sij T ) + λ i ( s ij d ij K i )) + λ n+1 ( r ij M (P f ) d ij P th (22) r )+λ n+2 ( (M d ij )δ co C th c )+λ n+3 ( (s ij d ij c sij T ) Cs th ) where K i = s ij d ij. The critical value of Λ is achieved only if its gradient is zero. Based on Theore 3.1, the NLIP optiization proble is convex. Thus, the gap between the relaxed proble and its dual proble is zero [4]. Also, the object function of the NLIP odel is derivable, thus the gradients of λ exist. We can get the solution for the optiization proble based on the Lagrange dual solution. Considering that the popularities and iportance of data objects usually do not change uch within a short period of tie, we can choose larger value for T to avoid coputation overhead, and also use IPOPT [41], [42] (a software library for large scale nonlinear optiization) to solve the large-scale nonlinear optiization proble, which ake ore practical. 4 IV. THE REPLICATION SCHEME In Section III, we calculate the replication degree of each data object, which is the first step of the design of. The next proble is how to assign the replicas to nodes, which is of iportance for increasing the data availability [9]. Recall that in Section III-A, we introduced BIBD-based file replication in Copyset Replication that can reduce data loss probability in correlated achine failures. Though the BIBDbased ethod can be used to reduce data loss probability in correlated achine failures, it requires a constant replication degree and cannot be used for replicating data with various replication degrees. Our proposed can deal with the probles. We present the details of below. Algorith 1 shows the pseudocode of the replication algorith. For particular and arbitrary i and j, the replication degree d ij of data D ij can be obtained fro the NLIP optiization odel in Section III. To reduce data loss in both correlated and non-correlated achine failures, first ranks these replication degrees in ascending order (i.e., d 1,...,d l ). For a given replication degree d i (i = 1,...,l), first groups the data objects with replication degree d i together (denoted by 3 The storage cost ay increase as storage space increases, but it also depends on the storage edia for storing the data, thus the clai that the space constraint holds at equality does not contradict the ain hypothesis (cost-effective) of the paper. 4 Although soe paraeters (e.g., data popularity, achine failure rate) are not usually available a priori, we can get the fro the historical data [7], [8], [16]. 247

7 Replication degree=d1 Replication degree=d2 Replication degree=d3 Server Server1 Server2 Server3 Server4 Server5 Server6 Server7 Server8 Server9 Server1 Server11 Server12 Server13 Server14 Server15 Server16 Server17 FTS 1 FTS 2 FTS 1 FTS 2 FTS 1 FTS 2 Server Server2 Server1 Server3 Server4 Server6 Server7 Server5 Server8 Server9 Server1 Server14 Server15 Server16 Server11 Server12 Server13 Server17 FTS 3 FTS 4 FTS 3 FTS 4 Table I: Paraeters fro publicly available data [12]. Syste Chunks per node Cluster size Scatter width Facebook HDFS RAMCloud N-1 D i ), and counts the nuber of data objects with replication degree d i (denoted by ND r i ). To handle the proble of varying replication degrees, partitions all nodes to l groups and then conducts Copyset Replication [12] to assign chunks with replication degree d i to the i th node group (i = 1,...,l). For load balance, the nuber of nodes in each group is Algorith 1: Pseudocode of the algorith. 1 Copute the replication degree for each data object using the NLIP odel (Section III) 2 Rank the replication degrees in ascending order d 1,...,d l 3 for i 1 to l do 4 Group data objects with d i together (D i ) 5 Use BIBD-based ethod to generate FTSs; each FTS has nodes fro different datacenters but within a certain geographic distance 6 Store each chunk s replicas of data objects with d i to all nodes in an FTS with d i 7 return proportional to the nuber of replicas that will be stored in the ND r d group. Accordingly, assigns N i i nodes to data l i=1 Nr D d i i group D i. then uses the BIBD-based ethod to generate FTSs for each group of nodes. Specifically, deterines λ in Section III-A1 based on the load balancing requireent and then uses the BIBD-based ethod for (N,R,λ), where N is the nuber of the nodes in a node group and R is the replication degree of the group (d i ) (R relates the correlated achine failure with replication degree). The nodes in each FTS are required to be fro different datacenters and within a certain geographic distance between each other. Distributing replicas over different datacenters can avoid data loss due to achine failures (e.g., caused by power outages) for data reliability. Liiting the distances between replica nodes for a data chunk constrains the consistency aintenance cost. replicates the chunks of each data object to all nodes in an FTS. 5 Figure 2 illustrates an exaple to show the design of. In the exaple, d 1 = 2, d 2 = 3, and d 3 = 4. In the left ost block, each FTS contains two chunks of a data object. Each FTS overlaps with every other FTS with at ost one server. This helps reduce the probability of data loss in correlated achine failures and balance the load. For those data objects with replication degree 2, the chunks of a data object will be stored to the nodes in an FTS in the left ost block. In the iddle block, each FTS contains three chunks of a data object. In the right ost block, each FTS contains four chunks of a data object. 5 Although putting all replicas of a chunk to the nodes in an FTS can bring about the cost of inter-rack transfer (across oversubscribed switches), it can significantly reduce the probability of data loss caused by correlated achine failures using BIBD-based ethod [12]. FTS 3 FTS 4 Figure 2: Architecture of for cloud storage fault-tolerant sets (FTSs). Table II: Paraeter settings. Paraeter Meaning Setting N # of servers 1-1 M # of chunks of a data object 3 [45] R # of servers in each FTS 3 λ # of FTSs containing a pair of servers 1 S Scatter width 4 p Prob. of a server failure.5 Pr th Threshold for expected request failure.5 Cc th Threshold for consistency aint. cost 1 Cs th Threshold for storage cost 3 # of data objects in each application 1 n # of data applications 5 Recall that each data priority value corresponds to a storage ediu. We assue that all servers have all types of storage edia, and we need to deterine which ediu to use for a given priority on a given server. When replicating a chunk to a node, chooses the storage edia for the chunks based on data objects priority calculated by Forula (13). Data objects with higher priority values will be stored to faster and ore expensive storage edia (e.g., Meory, SSD), and vice versa. The constraint (17) is to ensure that storage requireents of data objects do not overfill the servers. V. PERFORMANCE EVALUATION We conducted the nuerical analysis based on the paraeters in [12] (Table I) derived fro the syste statistics fro Facebook, HDFS and RAMCloud [1], [2], [1], [12], [13], [43], and also conducted real-world experients on Aazon S3. The distributions of file read and write rates in our analysis and tests follow those of the CTH trace [44] provided by Sandia National Laboratories that records 4-hour read/write log in a parallel file syste. A. Nuerical Analysis Results We copared with other three replication schees: Rando Replication (RR) [12], Replication Degree Custoization (RDC) [8] and Copyset Replication [12] (Copyset). We copare with Copyset instead of Tiered Replication (TR) [29] because TR focuses on data durability [29], however focus on data availability. Moreover, the key idea of Copyset and TR for reducing data loss probability is the sae, and they both use the BIBD-based ethod to reduce the probability of data loss. RR places the priary replica on a rando node (say node i) in the entire syste, and places the secondary replicas on (R-1) nodes around the priary node (i.e., nodes i+1, i+2,...). 6 RDC derives replication degree for each object with considering data object popularity to axiize the expected data availability for non-correlated achine failures. Copyset splits the nodes into a nuber of copysets, and constrains the replicas of every chunk to a single copyset so that it can reduce the frequency of 6 RR is based on Facebook s design, which chooses secondary replica holders fro a window of nodes around the priary node. 248

8 Probability of data loss Probability of data loss Probability of data loss Availability of requested data object (a) Facebook (b) HDFS (c) RAMCloud Figure 3: Probability of data loss vs. nuber of nodes (R = 3 for RR and Copyset). RDC Copyset RR Availability of requested data object RDC Copyset RR Availability of requested data object RDC Copyset RR Storage cost (a) Facebook (b) HDFS (c) RAMCloud Figure 4: Availability of requested data objects vs. nuber of nodes (R = 3 for RR and Copyset) Storage cost Copyset RDC RR 24 Copyset RDC RR 22 Storage cost Copyset RDC RR 22 5 (a) Facebook (b) HDFS (c) HDFS Figure 5: Storage cost vs. nuber of nodes (R = 3 for RR and Copyset). 5 5 Consistency aintenance cost Consistency aintenance cost Consistency aintenance cost (a) Facebook (b) HDFS (c) HDFS Figure 6: Consistency aintenance cost vs. nuber of nodes (R = 3 for RR and Copyset). data loss by iniizing the nuber of copysets. The nuber of nodes that fail concurrently in each syste was set to 1% of the nodes in the syste [6], [12]. Since this rate is the axiu percentage of concurrent failure nodes in real clouds (i.e., worst case) [1], [11], [46], it is reasonable to see higher probabilities of data loss and lower expected data availability in our analytical results than the real results in current clouds. The distributions of file popularity and updates follow those of the CTH trace. We used the noral distribution with ean of 1 and standard deviation of 1 to generate 1 unit costs for different storage edius. Copared to unifor achine failures, nonunifor and correlated achine failures are ore realistic due to different hardware/software copositions and configurations [8], [47]. Thus, we set w 1 =.4, w 2 =.2 and w 3 =.4 in Equ. (8), respectively. We randoly generated 6 bit nuber fro reasonable ranges for each node to represent its location, as explained in Section III-B. Table II shows the paraeter settings in our analysis unless otherwise specified. We first calculate the data loss probability for each ethod. 7 Specifically, we used Forula (8) for, Forula (2) for Copyset, Forula (3) for RR, and Equ. (8) with w 1 =, w 2 =.4 and w 3 =.6 for RDC. Figure 3(a)-3(c) show the 7 Many datacenter operators prefer to low probability of any incurring data loss at the expense of losing ore data in each event due to high cost of each incident of data loss [12] relationship between the probability of data loss and the nuber of nodes in the Facebook, HDFS and RAMCloud environents, respectively. We find that the result approxiately follows <Copyset<RDC<RR. The probability of data loss in Copyset is higher than that in. This is because considers non-correlated achine failures which are not considered in Copyset. RDC generates a higher probability of data loss than and Copyset because it neglects reducing the probability of data loss caused by correlated achine failures. The probability of data loss in RR is uch higher than and Copyset, and also higher than RDC. This is due to two reasons. First, RR places the copies of a chunk on a certain nuber (i.e., R) of different nodes. Any cobination of R nodes that fail siultaneously would result in data loss in correlated achine failures. However, and Copyset lose data only if all the nodes in an FTS fail siultaneously. Second, and RDC consider data popularity to increase the expected data availability, which is, however, not considered in RR. We then calculate the availability of requested data object by 1 P r, and P r is calculated by Forula (9). Figure 4(a)- 4(c) show the result of the availability of requested data objects. We see the result generally follows >Copyset>RDC>RR in all figures. The availability of requested data objects in Copyset is lower than that in because considers 249

9 Probability of data loss RR RDC Copyset Probability of data loss RR RDC Copyset (a) Scatter width=2 (b) Scatter width=4 Figure 7: Probability of data loss vs. nuber of nodes on Aazon S3 (R = 3 for RR and Copyset). data popularity when deterining the replication degree for each chunk, which is, however, not considered in Copyset. Also, reduces data loss in both correlated and noncorrelated achine failures, while Copyset only iniizes the data loss in correlated achine failures. The availability of requested data objects in Copyset is higher than that in RDC because RDC cannot reduce the probability of data loss caused by correlated achine failures and thereby decreases the availability of requested data objects. RR has the lowest availability because RR places the copies of a chunk on a certain nuber (i.e., R) of nodes and any cobination of R nodes that fail siultaneously would cause data loss. Also, RR does not consider data popularity as RDC. We then used Forula (15) to calculate the storage cost based on the sizes, replication degrees, and storage ediu unit costs of data objects for. For the other three ethods, we randoly choose storage edia for data objects and do not iniize the storage cost with space capacity constraint (Inequ. 17) for RDC. Figure 5(a)-5(c) show the result of storage cost. The result in all three figures generally follows RR Copyset>RDC>. This is because stores data objects into different storage edius based on their applications priorities, the sizes, and the replication degrees of data objects to iniize the total storage cost. The storage costs in Copyset and RR are higher than RDC and uch higher than. RDC reduces the replicas of unpopular data objects, thus reducing storage cost. Copyset and RR neither consider the different storage edius nor reduce the replicas of unpopular data objects to reduce storage cost. We used Forula (12) to calculate the consistency aintenance cost of each ethod based on the geographic distance and replication degrees of data objects. Figure 6(a)- 6(c) show the result of consistency aintenance cost. In these figures, the consistency aintenance costs in Copyset and RR are higher than because liits the geographic distance between the replica nodes of a chunk and the nuber of replicas, thereby reducing the consistency aintenance cost. However, Copyset and RR neglect geographic distances. RDC produces the highest consistency aintenance cost because RDC neglects geographic distance and it also generates ore replicas for popular data objects. B. Real-world Experiental Results We conducted experients on the real-world Aazon S3. We siulated the geo-distributed storage datacenters using three regions of Aazon S3 in the U.S. In each region, we created the sae nuber of buckets, each of which siulates a data server. The nuber of buckets was varied fro 5 to 25 with step size of 5 in our test. We distributed 5 Availability of requested data object RR Copyset RDC Availability of requested data object RR RDC Copyset (a) Scatter width=2 (b) Scatter width=4 Figure 8: Availability of requested data objects vs. nuber of nodes on Aazon S3 (R = 3 for RR and Copyset). Storage cost to Copyset 15% 1% 95% 9% 85% 8% 75% Copyset RDC RR Consistency aintenance cost to CopySet 12% 1% 8% 6% 4% 2% % (a) Storage cost (b) Consistency aintenance cost Figure 9: Storage cost and consistency aintenance cost vs. nuber of nodes on Aazon S3 (R= 3 for RR and Copyset). data objects to all the servers in the syste. We used the distributions of read and write rates fro the CTH trace data to generate reads and writes. The requests were generated fro servers in Windows Azure eastern region. We consider the requests targeting each region with latency ore than 1s as failed requests (unavailable data objects). 8 We used the paraeters in Table II except p and N. In this test, N is the total nuber of siulated data servers in the syste and p (with average value.89) follows the actual probability of a server failure in the real syste. We used the actual price of the data access of Aazon S3 to calculate the storage cost. Figure 7(a) and 7(b) show the result of probability of data loss on Aazon S3. We see the result approxiately follows <RDC<Copyset<RR. Our nuerical result shows that <Copyset<RDC<RR. Both results confir that generates the lowest probability of data loss. RDC generates higher probability of data loss than Copyset in the nuerical analysis but generates lower probability of data loss than Copyset in the experients. This is because RDC cannot handle correlated achine failures, and the failure rate of correlated achine failures is 1% in the nuerical analysis but our real-world experient has fewer correlated achine failures. We also see that scatter width 2 produces lower probability of data loss than scatter width 4. This is because a large scatter width increases the nuber of FTSs and thus increases the probability of data loss. Figure 8(a) and 8(b) show result of the availability of requested data objects on Aazon S3. We see the result follows >Copyset>RDC>RR, which is consistent with the result in Figure 4 due to the sae reasons. We also see that scatter width 2 produces higher availability than scatter width 4 because a large scatter width increases the probability of data loss and reduces data availability. We then regard Copyset as a baseline and calculate the ratio of storage/consistency aintenance cost of each of the other 8 Since it is hard to generate peranent failures on Aazon S3 and the network latency is low based on [48], we believe the data request failure within U.S. is ainly caused by achine failures, and we consider the requests targeting each region with latency longer than 1s as failed request, which reflects the availability of data objects. 25

Utility-Based Resource Allocation for Mixed Traffic in Wireless Networks

Utility-Based Resource Allocation for Mixed Traffic in Wireless Networks IEEE IFOCO 2 International Workshop on Future edia etworks and IP-based TV Utility-Based Resource Allocation for ixed Traffic in Wireless etworks Li Chen, Bin Wang, Xiaohang Chen, Xin Zhang, and Dacheng

More information

Energy-Efficient Disk Replacement and File Placement Techniques for Mobile Systems with Hard Disks

Energy-Efficient Disk Replacement and File Placement Techniques for Mobile Systems with Hard Disks Energy-Efficient Disk Replaceent and File Placeent Techniques for Mobile Systes with Hard Disks Young-Jin Ki School of Coputer Science & Engineering Seoul National University Seoul 151-742, KOREA youngjk@davinci.snu.ac.kr

More information

Solving the Damage Localization Problem in Structural Health Monitoring Using Techniques in Pattern Classification

Solving the Damage Localization Problem in Structural Health Monitoring Using Techniques in Pattern Classification Solving the Daage Localization Proble in Structural Health Monitoring Using Techniques in Pattern Classification CS 9 Final Project Due Dec. 4, 007 Hae Young Noh, Allen Cheung, Daxia Ge Introduction Structural

More information

Collaborative Web Caching Based on Proxy Affinities

Collaborative Web Caching Based on Proxy Affinities Collaborative Web Caching Based on Proxy Affinities Jiong Yang T J Watson Research Center IBM jiyang@usibco Wei Wang T J Watson Research Center IBM ww1@usibco Richard Muntz Coputer Science Departent UCLA

More information

Different criteria of dynamic routing

Different criteria of dynamic routing Procedia Coputer Science Volue 66, 2015, Pages 166 173 YSC 2015. 4th International Young Scientists Conference on Coputational Science Different criteria of dynaic routing Kurochkin 1*, Grinberg 1 1 Kharkevich

More information

Closing The Performance Gap between Causal Consistency and Eventual Consistency

Closing The Performance Gap between Causal Consistency and Eventual Consistency Closing The Perforance Gap between Causal Consistency and Eventual Consistency Jiaqing Du Călin Iorgulescu Aitabha Roy Willy Zwaenepoel EPFL ABSTRACT It is well known that causal consistency is ore expensive

More information

OPTIMAL COMPLEX SERVICES COMPOSITION IN SOA SYSTEMS

OPTIMAL COMPLEX SERVICES COMPOSITION IN SOA SYSTEMS Key words SOA, optial, coplex service, coposition, Quality of Service Piotr RYGIELSKI*, Paweł ŚWIĄTEK* OPTIMAL COMPLEX SERVICES COMPOSITION IN SOA SYSTEMS One of the ost iportant tasks in service oriented

More information

Performance Analysis of RAID in Different Workload

Performance Analysis of RAID in Different Workload Send Orders for Reprints to reprints@benthascience.ae 324 The Open Cybernetics & Systeics Journal, 2015, 9, 324-328 Perforance Analysis of RAID in Different Workload Open Access Zhang Dule *, Ji Xiaoyun,

More information

Clustering. Cluster Analysis of Microarray Data. Microarray Data for Clustering. Data for Clustering

Clustering. Cluster Analysis of Microarray Data. Microarray Data for Clustering. Data for Clustering Clustering Cluster Analysis of Microarray Data 4/3/009 Copyright 009 Dan Nettleton Group obects that are siilar to one another together in a cluster. Separate obects that are dissiilar fro each other into

More information

A Distributed Multi-service Resource Allocation Algorithm in Heterogeneous Wireless Access Medium

A Distributed Multi-service Resource Allocation Algorithm in Heterogeneous Wireless Access Medium 1 A Distributed Multi-service Resource Allocation Algorith in Heterogeneous Wireless Access Mediu Muhaad Isail, Student Meber, IEEE, and Weihua Zhuang, Fellow, IEEE Abstract In this paper, radio resource

More information

EE 364B Convex Optimization An ADMM Solution to the Sparse Coding Problem. Sonia Bhaskar, Will Zou Final Project Spring 2011

EE 364B Convex Optimization An ADMM Solution to the Sparse Coding Problem. Sonia Bhaskar, Will Zou Final Project Spring 2011 EE 364B Convex Optiization An ADMM Solution to the Sparse Coding Proble Sonia Bhaskar, Will Zou Final Project Spring 20 I. INTRODUCTION For our project, we apply the ethod of the alternating direction

More information

The optimization design of microphone array layout for wideband noise sources

The optimization design of microphone array layout for wideband noise sources PROCEEDINGS of the 22 nd International Congress on Acoustics Acoustic Array Systes: Paper ICA2016-903 The optiization design of icrophone array layout for wideband noise sources Pengxiao Teng (a), Jun

More information

A Fast Multi-Objective Genetic Algorithm for Hardware-Software Partitioning In Embedded System Design

A Fast Multi-Objective Genetic Algorithm for Hardware-Software Partitioning In Embedded System Design A Fast Multi-Obective Genetic Algorith for Hardware-Software Partitioning In Ebedded Syste Design 1 M.Jagadeeswari, 2 M.C.Bhuvaneswari 1 Research Scholar, P.S.G College of Technology, Coibatore, India

More information

An Efficient Approach for Content Delivery in Overlay Networks

An Efficient Approach for Content Delivery in Overlay Networks An Efficient Approach for Content Delivery in Overlay Networks Mohaad Malli, Chadi Barakat, Walid Dabbous Projet Planète, INRIA-Sophia Antipolis, France E-ail:{alli, cbarakat, dabbous}@sophia.inria.fr

More information

Identifying Converging Pairs of Nodes on a Budget

Identifying Converging Pairs of Nodes on a Budget Identifying Converging Pairs of Nodes on a Budget Konstantina Lazaridou Departent of Inforatics Aristotle University, Thessaloniki, Greece konlaznik@csd.auth.gr Evaggelia Pitoura Coputer Science and Engineering

More information

Adaptive Parameter Estimation Based Congestion Avoidance Strategy for DTN

Adaptive Parameter Estimation Based Congestion Avoidance Strategy for DTN Proceedings of the nd International onference on oputer Science and Electronics Engineering (ISEE 3) Adaptive Paraeter Estiation Based ongestion Avoidance Strategy for DTN Qicai Yang, Futong Qin, Jianquan

More information

Shortest Path Determination in a Wireless Packet Switch Network System in University of Calabar Using a Modified Dijkstra s Algorithm

Shortest Path Determination in a Wireless Packet Switch Network System in University of Calabar Using a Modified Dijkstra s Algorithm International Journal of Engineering and Technical Research (IJETR) ISSN: 31-869 (O) 454-4698 (P), Volue-5, Issue-1, May 16 Shortest Path Deterination in a Wireless Packet Switch Network Syste in University

More information

Oblivious Routing for Fat-Tree Based System Area Networks with Uncertain Traffic Demands

Oblivious Routing for Fat-Tree Based System Area Networks with Uncertain Traffic Demands Oblivious Routing for Fat-Tree Based Syste Area Networks with Uncertain Traffic Deands Xin Yuan Wickus Nienaber Zhenhai Duan Departent of Coputer Science Florida State University Tallahassee, FL 3306 {xyuan,nienaber,duan}@cs.fsu.edu

More information

Designing High Performance Web-Based Computing Services to Promote Telemedicine Database Management System

Designing High Performance Web-Based Computing Services to Promote Telemedicine Database Management System Designing High Perforance Web-Based Coputing Services to Proote Teleedicine Database Manageent Syste Isail Hababeh 1, Issa Khalil 2, and Abdallah Khreishah 3 1: Coputer Engineering & Inforation Technology,

More information

Fair Resource Allocation for Heterogeneous Tasks

Fair Resource Allocation for Heterogeneous Tasks Fair Resource Allocation for Heterogeneous Tasks Koyel Mukherjee, Partha utta, Gurulingesh Raravi, Thangaraj Balasubraania, Koustuv asgupta, Atul Singh Xerox Research Center India, Bangalore, India 560105

More information

Structural Balance in Networks. An Optimizational Approach. Andrej Mrvar. Faculty of Social Sciences. University of Ljubljana. Kardeljeva pl.

Structural Balance in Networks. An Optimizational Approach. Andrej Mrvar. Faculty of Social Sciences. University of Ljubljana. Kardeljeva pl. Structural Balance in Networks An Optiizational Approach Andrej Mrvar Faculty of Social Sciences University of Ljubljana Kardeljeva pl. 5 61109 Ljubljana March 23 1994 Contents 1 Balanced and clusterable

More information

A Hybrid Network Architecture for File Transfers

A Hybrid Network Architecture for File Transfers JOURNAL OF IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, VOL. 6, NO., JANUARY 9 A Hybrid Network Architecture for File Transfers Xiuduan Fang, Meber, IEEE, Malathi Veeraraghavan, Senior Meber,

More information

Survivability Function A Measure of Disaster-Based Routing Performance

Survivability Function A Measure of Disaster-Based Routing Performance Survivability Function A Measure of Disaster-Based Routing Perforance Journal Club Presentation on W. Molisz. Survivability function-a easure of disaster-based routing perforance. IEEE Journal on Selected

More information

Geo-activity Recommendations by using Improved Feature Combination

Geo-activity Recommendations by using Improved Feature Combination Geo-activity Recoendations by using Iproved Feature Cobination Masoud Sattari Middle East Technical University Ankara, Turkey e76326@ceng.etu.edu.tr Murat Manguoglu Middle East Technical University Ankara,

More information

Efficient Estimation of Inclusion Coefficient using HyperLogLog Sketches

Efficient Estimation of Inclusion Coefficient using HyperLogLog Sketches Efficient Estiation of Inclusion Coefficient using HyperLogLog Sketches Azade Nazi, Bolin Ding, Vivek Narasayya, Surajit Chaudhuri Microsoft Research {aznazi, bolind, viveknar, surajitc}@icrosoft.co ABSTRACT

More information

IN many interactive multimedia streaming applications, such

IN many interactive multimedia streaming applications, such 840 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 8, NO. 5, MAY 06 A Geoetric Approach to Server Selection for Interactive Video Streaing Yaochen Hu, Student Meber, IEEE, DiNiu, Meber, IEEE, and Zongpeng Li, Senior

More information

Using Imperialist Competitive Algorithm in Optimization of Nonlinear Multiple Responses

Using Imperialist Competitive Algorithm in Optimization of Nonlinear Multiple Responses International Journal of Industrial Engineering & Production Research Septeber 03, Volue 4, Nuber 3 pp. 9-35 ISSN: 008-4889 http://ijiepr.iust.ac.ir/ Using Iperialist Copetitive Algorith in Optiization

More information

A Periodic Dynamic Load Balancing Method

A Periodic Dynamic Load Balancing Method 2016 3 rd International Conference on Engineering Technology and Application (ICETA 2016) ISBN: 978-1-60595-383-0 A Periodic Dynaic Load Balancing Method Taotao Fan* State Key Laboratory of Inforation

More information

Data & Knowledge Engineering

Data & Knowledge Engineering Data & Knowledge Engineering 7 (211) 17 187 Contents lists available at ScienceDirect Data & Knowledge Engineering journal hoepage: www.elsevier.co/locate/datak An approxiate duplicate eliination in RFID

More information

THE rapid growth and continuous change of the real

THE rapid growth and continuous change of the real IEEE TRANSACTIONS ON SERVICES COMPUTING, VOL. 8, NO. 1, JANUARY/FEBRUARY 2015 47 Designing High Perforance Web-Based Coputing Services to Proote Teleedicine Database Manageent Syste Isail Hababeh, Issa

More information

Fair Energy-Efficient Sensing Task Allocation in Participatory Sensing with Smartphones

Fair Energy-Efficient Sensing Task Allocation in Participatory Sensing with Smartphones Advance Access publication on 24 February 2017 The British Coputer Society 2017. All rights reserved. For Perissions, please eail: journals.perissions@oup.co doi:10.1093/cojnl/bxx015 Fair Energy-Efficient

More information

Generating Mechanisms for Evolving Software Mirror Graph

Generating Mechanisms for Evolving Software Mirror Graph Journal of Modern Physics, 2012, 3, 1050-1059 http://dx.doi.org/10.4236/jp.2012.39139 Published Online Septeber 2012 (http://www.scirp.org/journal/jp) Generating Mechaniss for Evolving Software Mirror

More information

Multipath Selection and Channel Assignment in Wireless Mesh Networks

Multipath Selection and Channel Assignment in Wireless Mesh Networks Multipath Selection and Channel Assignent in Wireless Mesh Networs Soo-young Jang and Chae Y. Lee Dept. of Industrial and Systes Engineering, KAIST, 373-1 Kusung-dong, Taejon, Korea Tel: +82-42-350-5916,

More information

QUERY ROUTING OPTIMIZATION IN SENSOR COMMUNICATION NETWORKS

QUERY ROUTING OPTIMIZATION IN SENSOR COMMUNICATION NETWORKS QUERY ROUTING OPTIMIZATION IN SENSOR COMMUNICATION NETWORKS Guofei Jiang and George Cybenko Institute for Security Technology Studies and Thayer School of Engineering Dartouth College, Hanover NH 03755

More information

A Novel Fast Constructive Algorithm for Neural Classifier

A Novel Fast Constructive Algorithm for Neural Classifier A Novel Fast Constructive Algorith for Neural Classifier Xudong Jiang Centre for Signal Processing, School of Electrical and Electronic Engineering Nanyang Technological University Nanyang Avenue, Singapore

More information

Modeling Parallel Applications Performance on Heterogeneous Systems

Modeling Parallel Applications Performance on Heterogeneous Systems Modeling Parallel Applications Perforance on Heterogeneous Systes Jaeela Al-Jaroodi, Nader Mohaed, Hong Jiang and David Swanson Departent of Coputer Science and Engineering University of Nebraska Lincoln

More information

A Comparative Study of Two-phase Heuristic Approaches to General Job Shop Scheduling Problem

A Comparative Study of Two-phase Heuristic Approaches to General Job Shop Scheduling Problem IE Vol. 7, No. 2, pp. 84-92, Septeber 2008. A Coparative Study of Two-phase Heuristic Approaches to General Job Shop Scheduling Proble Ji Ung Sun School of Industrial and Managent Engineering Hankuk University

More information

Defining and Surveying Wireless Link Virtualization and Wireless Network Virtualization

Defining and Surveying Wireless Link Virtualization and Wireless Network Virtualization 1 Defining and Surveying Wireless Link Virtualization and Wireless Network Virtualization Jonathan van de Belt, Haed Ahadi, and Linda E. Doyle The Centre for Future Networks and Counications - CONNECT,

More information

Ravine Streams: Persistent Data Streams in Disruptive Sensor Networks

Ravine Streams: Persistent Data Streams in Disruptive Sensor Networks Ravine Streas: Persistent Data Streas in Disruptive Sensor Networks Mingsen Xu, Wen-Zhan Song Departent of Coputer Science, Georgia State University Atlanta, Georgia 333, USA Eail: {xu4@student.gsu.edu,

More information

Planet Scale Software Updates

Planet Scale Software Updates Planet Scale Software Updates Christos Gkantsidis 1, Thoas Karagiannis 2, Pablo Rodriguez 1, and Milan Vojnović 1 1 Microsoft Research 2 UC Riverside Cabridge, UK Riverside, CA, USA {chrisgk,pablo,ilanv}@icrosoft.co

More information

Design Optimization of Mixed Time/Event-Triggered Distributed Embedded Systems

Design Optimization of Mixed Time/Event-Triggered Distributed Embedded Systems Design Optiization of Mixed Tie/Event-Triggered Distributed Ebedded Systes Traian Pop, Petru Eles, Zebo Peng Dept. of Coputer and Inforation Science, Linköping University {trapo, petel, zebpe}@ida.liu.se

More information

Derivation of an Analytical Model for Evaluating the Performance of a Multi- Queue Nodes Network Router

Derivation of an Analytical Model for Evaluating the Performance of a Multi- Queue Nodes Network Router Derivation of an Analytical Model for Evaluating the Perforance of a Multi- Queue Nodes Network Router 1 Hussein Al-Bahadili, 1 Jafar Ababneh, and 2 Fadi Thabtah 1 Coputer Inforation Systes Faculty of

More information

Knowledge Discovery Applied to Agriculture Economy Planning

Knowledge Discovery Applied to Agriculture Economy Planning Knowledge Discovery Applied to Agriculture Econoy Planning Bing-ru Yang and Shao-un Huang Inforation Engineering School University of Science and Technology, Beiing, China, 100083 Eail: bingru.yang@b.col.co.cn

More information

Scheduling Parallel Real-Time Recurrent Tasks on Multicore Platforms

Scheduling Parallel Real-Time Recurrent Tasks on Multicore Platforms IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, VOL., NO., NOV 27 Scheduling Parallel Real-Tie Recurrent Tasks on Multicore Platfors Risat Pathan, Petros Voudouris, and Per Stenströ Abstract We

More information

Gromov-Hausdorff Distance Between Metric Graphs

Gromov-Hausdorff Distance Between Metric Graphs Groov-Hausdorff Distance Between Metric Graphs Jiwon Choi St Mark s School January, 019 Abstract In this paper we study the Groov-Hausdorff distance between two etric graphs We copute the precise value

More information

Scheduling Parallel Task Graphs on (Almost) Homogeneous Multi-cluster Platforms

Scheduling Parallel Task Graphs on (Almost) Homogeneous Multi-cluster Platforms 1 Scheduling Parallel Task Graphs on (Alost) Hoogeneous Multi-cluster Platfors Pierre-François Dutot, Tchiou N Takpé, Frédéric Suter and Henri Casanova Abstract Applications structured as parallel task

More information

Research on a Kind of QoS-Sensitive Semantic Web Services Composition Method Based on Genetic Algorithm

Research on a Kind of QoS-Sensitive Semantic Web Services Composition Method Based on Genetic Algorithm roceedings of the 7th International Conference on Innovation & Manageent 893 Research on a Kind of QoS-Sensitive Seantic Web Services Coposition Method Based on Genetic Algorith Cao Hongjiang, Nie Guihua,

More information

Detection of Outliers and Reduction of their Undesirable Effects for Improving the Accuracy of K-means Clustering Algorithm

Detection of Outliers and Reduction of their Undesirable Effects for Improving the Accuracy of K-means Clustering Algorithm Detection of Outliers and Reduction of their Undesirable Effects for Iproving the Accuracy of K-eans Clustering Algorith Bahan Askari Departent of Coputer Science and Research Branch, Islaic Azad University,

More information

A Beam Search Method to Solve the Problem of Assignment Cells to Switches in a Cellular Mobile Network

A Beam Search Method to Solve the Problem of Assignment Cells to Switches in a Cellular Mobile Network A Bea Search Method to Solve the Proble of Assignent Cells to Switches in a Cellular Mobile Networ Cassilda Maria Ribeiro Faculdade de Engenharia de Guaratinguetá - DMA UNESP - São Paulo State University

More information

Mapping Data in Peer-to-Peer Systems: Semantics and Algorithmic Issues

Mapping Data in Peer-to-Peer Systems: Semantics and Algorithmic Issues Mapping Data in Peer-to-Peer Systes: Seantics and Algorithic Issues Anastasios Keentsietsidis Marcelo Arenas Renée J. Miller Departent of Coputer Science University of Toronto {tasos,arenas,iller}@cs.toronto.edu

More information

Investigation of The Time-Offset-Based QoS Support with Optical Burst Switching in WDM Networks

Investigation of The Time-Offset-Based QoS Support with Optical Burst Switching in WDM Networks Investigation of The Tie-Offset-Based QoS Support with Optical Burst Switching in WDM Networks Pingyi Fan, Chongxi Feng,Yichao Wang, Ning Ge State Key Laboratory on Microwave and Digital Counications,

More information

Automatic Graph Drawing Algorithms

Automatic Graph Drawing Algorithms Autoatic Graph Drawing Algoriths Susan Si sisuz@turing.utoronto.ca Deceber 7, 996. Ebeddings of graphs have been of interest to theoreticians for soe tie, in particular those of planar graphs and graphs

More information

Improve Peer Cooperation using Social Networks

Improve Peer Cooperation using Social Networks Iprove Peer Cooperation using Social Networks Victor Ponce, Jie Wu, and Xiuqi Li Departent of Coputer Science and Engineering Florida Atlantic University Boca Raton, FL 33431 Noveber 5, 2007 Corresponding

More information

Redundancy Level Impact of the Mean Time to Failure on Wireless Sensor Network

Redundancy Level Impact of the Mean Time to Failure on Wireless Sensor Network (IJACSA) International Journal of Advanced Coputer Science and Applications Vol. 8, No. 1, 217 Redundancy Level Ipact of the Mean Tie to Failure on Wireless Sensor Network Alaa E. S. Ahed 1 College of

More information

Research Article A Game Theory Based Congestion Control Protocol for Wireless Personal Area Networks

Research Article A Game Theory Based Congestion Control Protocol for Wireless Personal Area Networks Journal of Sensors Volue 216, Article ID 61683, 13 pages http://dx.doi.org/1.11/216/61683 Research Article A Gae Theory Based Congestion Control Protocol for Wireless Personal Area Networks Chuang Ma,

More information

Evaluation of a multi-frame blind deconvolution algorithm using Cramér-Rao bounds

Evaluation of a multi-frame blind deconvolution algorithm using Cramér-Rao bounds Evaluation of a ulti-frae blind deconvolution algorith using Craér-Rao bounds Charles C. Beckner, Jr. Air Force Research Laboratory, 3550 Aberdeen Ave SE, Kirtland AFB, New Mexico, USA 87117-5776 Charles

More information

Content-Centric Multicast Beamforming in Cache-Enabled Cloud Radio Access Networks

Content-Centric Multicast Beamforming in Cache-Enabled Cloud Radio Access Networks Content-Centric Multicast Beaforing in Cache-Enabled Cloud Radio Access Networks Hao Zhou, Meixia Tao, Erkai Chen, Wei Yu *Dept. of Electronic Engineering, Shanghai Jiao Tong University, Shanghai, China

More information

Distributed Multicast Tree Construction in Wireless Sensor Networks

Distributed Multicast Tree Construction in Wireless Sensor Networks Distributed Multicast Tree Construction in Wireless Sensor Networks Hongyu Gong, Luoyi Fu, Xinzhe Fu, Lutian Zhao 3, Kainan Wang, and Xinbing Wang Dept. of Electronic Engineering, Shanghai Jiao Tong University,

More information

Integrating fast mobility in the OLSR routing protocol

Integrating fast mobility in the OLSR routing protocol Integrating fast obility in the OLSR routing protocol Mounir BENZAID 1,2, Pascale MINET 1 and Khaldoun AL AGHA 1,2 1 INRIA, Doaine de Voluceau - B.P.105, 78153 Le Chesnay Cedex, FRANCE ounir.benzaid, pascale.inet@inria.fr

More information

TensorFlow and Keras-based Convolutional Neural Network in CAT Image Recognition Ang LI 1,*, Yi-xiang LI 2 and Xue-hui LI 3

TensorFlow and Keras-based Convolutional Neural Network in CAT Image Recognition Ang LI 1,*, Yi-xiang LI 2 and Xue-hui LI 3 2017 2nd International Conference on Coputational Modeling, Siulation and Applied Matheatics (CMSAM 2017) ISBN: 978-1-60595-499-8 TensorFlow and Keras-based Convolutional Neural Network in CAT Iage Recognition

More information

I ve Seen Enough : Incrementally Improving Visualizations to Support Rapid Decision Making

I ve Seen Enough : Incrementally Improving Visualizations to Support Rapid Decision Making I ve Seen Enough : Increentally Iproving Visualizations to Support Rapid Decision Making Sajjadur Rahan Marya Aliakbarpour 2 Ha Kyung Kong Eric Blais 3 Karrie Karahalios,4 Aditya Paraeswaran Ronitt Rubinfield

More information

Theoretical Analysis of Local Search and Simple Evolutionary Algorithms for the Generalized Travelling Salesperson Problem

Theoretical Analysis of Local Search and Simple Evolutionary Algorithms for the Generalized Travelling Salesperson Problem Theoretical Analysis of Local Search and Siple Evolutionary Algoriths for the Generalized Travelling Salesperson Proble Mojgan Pourhassan ojgan.pourhassan@adelaide.edu.au Optiisation and Logistics, The

More information

Resource Optimization for Web Service Composition

Resource Optimization for Web Service Composition Resource Optiization for Web Coposition Xia Gao, Ravi Jain, Zulfikar Razan, Ulas Kozat Abstract coposition recently eerged as a costeffective way to quickly create new services within a network. Soe research

More information

Adaptive Holistic Scheduling for In-Network Sensor Query Processing

Adaptive Holistic Scheduling for In-Network Sensor Query Processing Adaptive Holistic Scheduling for In-Network Sensor Query Processing Hejun Wu and Qiong Luo Departent of Coputer Science and Engineering Hong Kong University of Science & Technology Clear Water Bay, Kowloon,

More information

Secure Wireless Multihop Transmissions by Intentional Collisions with Noise Wireless Signals

Secure Wireless Multihop Transmissions by Intentional Collisions with Noise Wireless Signals Int'l Conf. Wireless etworks ICW'16 51 Secure Wireless Multihop Transissions by Intentional Collisions with oise Wireless Signals Isau Shiada 1 and Hiroaki Higaki 1 1 Tokyo Denki University, Japan Abstract

More information

1 Extended Boolean Model

1 Extended Boolean Model 1 EXTENDED BOOLEAN MODEL It has been well-known that the Boolean odel is too inflexible, requiring skilful use of Boolean operators to obtain good results. On the other hand, the vector space odel is flexible

More information

Geometry. The Method of the Center of Mass (mass points): Solving problems using the Law of Lever (mass points). Menelaus theorem. Pappus theorem.

Geometry. The Method of the Center of Mass (mass points): Solving problems using the Law of Lever (mass points). Menelaus theorem. Pappus theorem. Noveber 13, 2016 Geoetry. The Method of the enter of Mass (ass points): Solving probles using the Law of Lever (ass points). Menelaus theore. Pappus theore. M d Theore (Law of Lever). Masses (weights)

More information

Predicting x86 Program Runtime for Intel Processor

Predicting x86 Program Runtime for Intel Processor Predicting x86 Progra Runtie for Intel Processor Behra Mistree, Haidreza Haki Javadi, Oid Mashayekhi Stanford University bistree,hrhaki,oid@stanford.edu 1 Introduction Progras can be equivalent: given

More information

COLOR HISTOGRAM AND DISCRETE COSINE TRANSFORM FOR COLOR IMAGE RETRIEVAL

COLOR HISTOGRAM AND DISCRETE COSINE TRANSFORM FOR COLOR IMAGE RETRIEVAL COLOR HISTOGRAM AND DISCRETE COSINE TRANSFORM FOR COLOR IMAGE RETRIEVAL 1 Te-Wei Chiang ( 蔣德威 ), 2 Tienwei Tsai ( 蔡殿偉 ), 3 Jeng-Ping Lin ( 林正平 ) 1 Dept. of Accounting Inforation Systes, Chilee Institute

More information

Copysets: Reducing the Frequency of Data Loss in Cloud Storage

Copysets: Reducing the Frequency of Data Loss in Cloud Storage Copysets: Reducing the Frequency of Data Loss in Cloud Storage Asaf Cidon, Stephen M. Rumble, Ryan Stutsman, Sachin Katti, John Ousterhout and Mendel Rosenblum Stanford University cidon@stanford.edu, {rumble,stutsman,skatti,ouster,mendel}@cs.stanford.edu

More information

BP-MAC: A High Reliable Backoff Preamble MAC Protocol for Wireless Sensor Networks

BP-MAC: A High Reliable Backoff Preamble MAC Protocol for Wireless Sensor Networks BP-MAC: A High Reliable Backoff Preable MAC Protocol for Wireless Sensor Networks Alexander Klein* Jirka Klaue Josef Schalk EADS Innovation Works, Gerany *Eail: klein@inforatik.uni-wuerzburg.de ABSTRACT:

More information

Database Design on Mechanical Equipment Operation Management System Zheng Qiu1, Wu kaiyuan1, Wu Chengyan1, Liu Lei2

Database Design on Mechanical Equipment Operation Management System Zheng Qiu1, Wu kaiyuan1, Wu Chengyan1, Liu Lei2 2nd International Conference on Advances in Mechanical Engineering and Industrial Inforatics (AMEII 206) Database Design on Mechanical Equipent Manageent Syste Zheng Qiu, Wu kaiyuan, Wu Chengyan, Liu Lei2

More information

Evaluation of the Timing Properties of Two Control Networks: CAN and PROFIBUS

Evaluation of the Timing Properties of Two Control Networks: CAN and PROFIBUS Evaluation of the Tiing Properties of Two Control Networs: CAN and PROFIBUS Max Mauro Dias Santos 1, Marcelo Ricardo Steer 2 and Francisco Vasques 3 1 UnilesteMG, CEP 35170-056, Coronel Fabriciano MG Brasil.

More information

TALLINN UNIVERSITY OF TECHNOLOGY, INSTITUTE OF PHYSICS 17. FRESNEL DIFFRACTION ON A ROUND APERTURE

TALLINN UNIVERSITY OF TECHNOLOGY, INSTITUTE OF PHYSICS 17. FRESNEL DIFFRACTION ON A ROUND APERTURE 7. FRESNEL DIFFRACTION ON A ROUND APERTURE. Objective Exaining diffraction pattern on a round aperture, deterining wavelength of light source.. Equipent needed Optical workbench, light source, color filters,

More information

Privacy-preserving String-Matching With PRAM Algorithms

Privacy-preserving String-Matching With PRAM Algorithms Privacy-preserving String-Matching With PRAM Algoriths Report in MTAT.07.022 Research Seinar in Cryptography, Fall 2014 Author: Sander Sii Supervisor: Peeter Laud Deceber 14, 2014 Abstract In this report,

More information

A Trajectory Splitting Model for Efficient Spatio-Temporal Indexing

A Trajectory Splitting Model for Efficient Spatio-Temporal Indexing A Trajectory Splitting Model for Efficient Spatio-Teporal Indexing Slobodan Rasetic Jörg Sander Jaes Elding Mario A. Nasciento Departent of Coputing Science University of Alberta Edonton, Alberta, Canada

More information

HKBU Institutional Repository

HKBU Institutional Repository Hong Kong Baptist University HKBU Institutional Repository HKBU Staff Publication 18 Towards why-not spatial keyword top-k ueries: A direction-aware approach Lei Chen Hong Kong Baptist University, psuanle@gail.co

More information

An Integrated Processing Method for Multiple Large-scale Point-Clouds Captured from Different Viewpoints

An Integrated Processing Method for Multiple Large-scale Point-Clouds Captured from Different Viewpoints 519 An Integrated Processing Method for Multiple Large-scale Point-Clouds Captured fro Different Viewpoints Yousuke Kawauchi 1, Shin Usuki, Kenjiro T. Miura 3, Hiroshi Masuda 4 and Ichiro Tanaka 5 1 Shizuoka

More information

Feature Selection to Relate Words and Images

Feature Selection to Relate Words and Images The Open Inforation Systes Journal, 2009, 3, 9-13 9 Feature Selection to Relate Words and Iages Wei-Chao Lin 1 and Chih-Fong Tsai*,2 Open Access 1 Departent of Coputing, Engineering and Technology, University

More information

Optimal Route Queries with Arbitrary Order Constraints

Optimal Route Queries with Arbitrary Order Constraints IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL.?, NO.?,? 20?? 1 Optial Route Queries with Arbitrary Order Constraints Jing Li, Yin Yang, Nikos Maoulis Abstract Given a set of spatial points DS,

More information

The Application of Bandwidth Optimization Technique in SLA Negotiation Process

The Application of Bandwidth Optimization Technique in SLA Negotiation Process The Application of Bandwidth Optiization Technique in SLA egotiation Process Srećko Krile University of Dubrovnik Departent of Electrical Engineering and Coputing Cira Carica 4, 20000 Dubrovnik, Croatia

More information

Real-Time Detection of Invisible Spreaders

Real-Time Detection of Invisible Spreaders Real-Tie Detection of Invisible Spreaders MyungKeun Yoon Shigang Chen Departent of Coputer & Inforation Science & Engineering University of Florida, Gainesville, FL 3, USA {yoon, sgchen}@cise.ufl.edu Abstract

More information

International Journal of Scientific & Engineering Research, Volume 4, Issue 10, October ISSN

International Journal of Scientific & Engineering Research, Volume 4, Issue 10, October ISSN International Journal of Scientific & Engineering Research, Volue 4, Issue 0, October-203 483 Design an Encoding Technique Using Forbidden Transition free Algorith to Reduce Cross-Talk for On-Chip VLSI

More information

RECONFIGURABLE AND MODULAR BASED SYNTHESIS OF CYCLIC DSP DATA FLOW GRAPHS

RECONFIGURABLE AND MODULAR BASED SYNTHESIS OF CYCLIC DSP DATA FLOW GRAPHS RECONFIGURABLE AND MODULAR BASED SYNTHESIS OF CYCLIC DSP DATA FLOW GRAPHS AWNI ITRADAT Assistant Professor, Departent of Coputer Engineering, Faculty of Engineering, Hasheite University, P.O. Box 15459,

More information

A Broadband Spectrum Sensing Algorithm in TDCS Based on ICoSaMP Reconstruction

A Broadband Spectrum Sensing Algorithm in TDCS Based on ICoSaMP Reconstruction MATEC Web of Conferences 73, 03073(08) https://doi.org/0.05/atecconf/087303073 SMIMA 08 A roadband Spectru Sensing Algorith in TDCS ased on I Reconstruction Liu Yang, Ren Qinghua, Xu ingzheng and Li Xiazhao

More information

A Measurement-Based Model for Parallel Real-Time Tasks

A Measurement-Based Model for Parallel Real-Time Tasks A Measureent-Based Model for Parallel Real-Tie Tasks Kunal Agrawal 1 Washington University in St. Louis St. Louis, MO, USA kunal@wustl.edu https://orcid.org/0000-0001-5882-6647 Sanjoy Baruah 2 Washington

More information

Enhancing Real-Time CAN Communications by the Prioritization of Urgent Messages at the Outgoing Queue

Enhancing Real-Time CAN Communications by the Prioritization of Urgent Messages at the Outgoing Queue Enhancing Real-Tie CAN Counications by the Prioritization of Urgent Messages at the Outgoing Queue ANTÓNIO J. PIRES (1), JOÃO P. SOUSA (), FRANCISCO VASQUES (3) 1,,3 Faculdade de Engenharia da Universidade

More information

A Discrete Spring Model to Generate Fair Curves and Surfaces

A Discrete Spring Model to Generate Fair Curves and Surfaces A Discrete Spring Model to Generate Fair Curves and Surfaces Atsushi Yaada Kenji Shiada 2 Tootake Furuhata and Ko-Hsiu Hou 2 Tokyo Research Laboratory IBM Japan Ltd. LAB-S73 623-4 Shiotsurua Yaato Kanagawa

More information

QoS and Sensible Routing Decisions

QoS and Sensible Routing Decisions QoS and Sensible Routing Decisions Erol Gelenbe Dept. of Electrical & Electronic Engineering Iperial College London SW7 2BT e.gelenbe@iperial.ac.uk Abstract Network Quality of Service (QoS) criteria of

More information

Handling Indivisibilities. Notes for AGEC 622. Bruce McCarl Regents Professor of Agricultural Economics Texas A&M University

Handling Indivisibilities. Notes for AGEC 622. Bruce McCarl Regents Professor of Agricultural Economics Texas A&M University Notes for AGEC 6 Bruce McCarl Regents Professor of Agricultural Econoics Texas A&M University All or None -- Integer Prograing McCarl and Spreen Chapter 5 Many investent and probles involve cases where

More information

Joint Measurement- and Traffic Descriptor-based Admission Control at Real-Time Traffic Aggregation Points

Joint Measurement- and Traffic Descriptor-based Admission Control at Real-Time Traffic Aggregation Points Joint Measureent- and Traffic Descriptor-based Adission Control at Real-Tie Traffic Aggregation Points Stylianos Georgoulas, Panos Triintzios and George Pavlou Centre for Counication Systes Research, University

More information

Dynamical Topology Analysis of VANET Based on Complex Networks Theory

Dynamical Topology Analysis of VANET Based on Complex Networks Theory BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volue 14, Special Issue Sofia 014 Print ISSN: 1311-970; Online ISSN: 1314-4081 DOI: 10.478/cait-014-0053 Dynaical Topology Analysis

More information

Collection Selection Based on Historical Performance for Efficient Processing

Collection Selection Based on Historical Performance for Efficient Processing Collection Selection Based on Historical Perforance for Efficient Processing Christopher T. Fallen and Gregory B. Newby Arctic Region Supercoputing Center University of Alaska Fairbanks Fairbanks, Alaska

More information

Deterministic Voting in Distributed Systems Using Error-Correcting Codes

Deterministic Voting in Distributed Systems Using Error-Correcting Codes IEEE TRASACTIOS O PARALLEL AD DISTRIBUTED SYSTEMS, VOL. 9, O. 8, AUGUST 1998 813 Deterinistic Voting in Distributed Systes Using Error-Correcting Codes Lihao Xu and Jehoshua Bruck, Senior Meber, IEEE Abstract

More information

Leveraging Relevance Cues for Improved Spoken Document Retrieval

Leveraging Relevance Cues for Improved Spoken Document Retrieval Leveraging Relevance Cues for Iproved Spoken Docuent Retrieval Pei-Ning Chen 1, Kuan-Yu Chen 2 and Berlin Chen 1 National Taiwan Noral University, Taiwan 1 Institute of Inforation Science, Acadeia Sinica,

More information

An Ad Hoc Adaptive Hashing Technique for Non-Uniformly Distributed IP Address Lookup in Computer Networks

An Ad Hoc Adaptive Hashing Technique for Non-Uniformly Distributed IP Address Lookup in Computer Networks An Ad Hoc Adaptive Hashing Technique for Non-Uniforly Distributed IP Address Lookup in Coputer Networks Christopher Martinez Departent of Electrical and Coputer Engineering The University of Texas at San

More information

Balancing Traffic Load for Devolved Controllers in Data Center Networks

Balancing Traffic Load for Devolved Controllers in Data Center Networks Globeco - Syposiu on Selected Areas in Counications: GC SAC Cloud Networks Balancing Traffic Load for Devolved Controllers in Data Center Networks Wanchao Liang, Xiaofeng Gao, Fan Wu, Guihai Chen, Wei

More information

Rule Extraction using Artificial Neural Networks

Rule Extraction using Artificial Neural Networks Rule Extraction using Artificial Neural Networks S. M. Karuzzaan 1 Ahed Ryadh Hasan 2 Abstract Artificial neural networks have been successfully applied to a variety of business application probles involving

More information

Boosted Detection of Objects and Attributes

Boosted Detection of Objects and Attributes L M M Boosted Detection of Objects and Attributes Abstract We present a new fraework for detection of object and attributes in iages based on boosted cobination of priitive classifiers. The fraework directly

More information

A Low-cost Memory Architecture with NAND XIP for Mobile Embedded Systems

A Low-cost Memory Architecture with NAND XIP for Mobile Embedded Systems A Low-cost Meory Architecture with XIP for Mobile Ebedded Systes Chanik Park, Jaeyu Seo, Sunghwan Bae, Hyojun Ki, Shinhan Ki and Busoo Ki Software Center, SAMSUNG Electronics, Co., Ltd. Seoul 135-893,

More information