Deadline Guaranteed Service for Multi- Tenant Cloud Storage Guoxin Liu and Haiying Shen

Similar documents
Towards Deadline Guaranteed Cloud Storage Services Guoxin Liu, Haiying Shen, and Lei Yu

Minimum-cost Cloud Storage Service Across Multiple Cloud Providers

An Economical and SLO- Guaranteed Cloud Storage Service across Multiple Cloud Service Providers Guoxin Liu and Haiying Shen

Better Never than Late: Meeting Deadlines in Datacenter Networks

Towards Deadline Guaranteed Cloud Storage Services

Consolidating Complementary VMs with Spatial/Temporalawareness

Shen, Tang, Yang, and Chu

Minimum-cost Cloud Storage Service Across Multiple Cloud Providers

Venice: Reliable Virtual Data Center Embedding in Clouds

Nested QoS: Providing Flexible Performance in Shared IO Environment

Optimizing Network Performance in Distributed Machine Learning. Luo Mai Chuntao Hong Paolo Costa

MODELING OF SMART GRID TRAFFICS USING NON- PREEMPTIVE PRIORITY QUEUES

Selective Data Replication for Online Social Networks with Distributed Datacenters Guoxin Liu *, Haiying Shen *, Harrison Chandler *

Pricing Intra-Datacenter Networks with

MicroFuge: A Middleware Approach to Providing Performance Isolation in Cloud Storage Systems

DDSS: Dynamic Dedicated Servers Scheduling for Multi Priority Level Classes in Cloud Computing

Quality-Assured Cloud Bandwidth Auto-Scaling for Video-on-Demand Applications

Hardware Evolution in Data Centers

CHAPTER 6 STATISTICAL MODELING OF REAL WORLD CLOUD ENVIRONMENT FOR RELIABILITY AND ITS EFFECT ON ENERGY AND PERFORMANCE

Distributed Autonomous Virtual Resource Management in Datacenters Using Finite- Markov Decision Process

Elastic Virtual Network Function Placement CloudNet 2015

Application Placement and Demand Distribution in a Global Elastic Cloud: A Unified Approach

An Efficient Holistic Data Distribution and Storage Solution for Online Social Networks

IX: A Protected Dataplane Operating System for High Throughput and Low Latency

Priority-Aware Virtual Machine Selection Algorithm in Dynamic Consolidation

Performance Assurance in Virtualized Data Centers

On-Demand Bandwidth Pricing for Congestion Control in Core Switches in Cloud Networks

Pocket: Elastic Ephemeral Storage for Serverless Analytics

DIAL: A Distributed Adaptive-Learning Routing Method in VDTNs

ReViNE: Reallocation of Virtual Network Embedding to Eliminate Substrate Bottleneck

CQNCR: Optimal VM Migration Planning in Cloud Data Centers

HPC in Cloud. Presenter: Naresh K. Sehgal Contributors: Billy Cox, John M. Acken, Sohum Sohoni

Basics (cont.) Characteristics of data communication technologies OSI-Model

Telecommunication Services Engineering Lab. Roch H. Glitho

Estimate performance and capacity requirements for Access Services

Towards Green Cloud Computing: Demand Allocation and Pricing Policies for Cloud Service Brokerage Chenxi Qiu

Dynamically Provisioning Distributed Systems to Meet Target Levels of Performance, Availability, and Data Quality

Optimizing SAN Performance & Availability: A Financial Services Customer Perspective

Energy Aware Scheduling in Cloud Datacenter

Optical Packet Switching

Femto-Matching: Efficient Traffic Offloading in Heterogeneous Cellular Networks

Week 7: Traffic Models and QoS

8. CONCLUSION AND FUTURE WORK. To address the formulated research issues, this thesis has achieved each of the objectives delineated in Chapter 1.

A QoS Load Balancing Scheduling Algorithm in Cloud Environment

A Network-aware Scheduler in Data-parallel Clusters for High Performance

Considering Resource Demand Misalignments To Reduce Resource Over-Provisioning in Cloud Datacenters

Network Support for Multimedia

Episode 5. Scheduling and Traffic Management

Master Course Computer Networks IN2097

Priority Traffic CSCD 433/533. Advanced Networks Spring Lecture 21 Congestion Control and Queuing Strategies

Efficient Point to Multipoint Transfers Across Datacenters

Call Admission Control in IP networks with QoS support

Master Course Computer Networks IN2097

Power-Aware Virtual Machine Scheduling-policy for Virtualized Heterogeneous Multicore Systems

No Tradeoff Low Latency + High Efficiency

Dynamic Virtual Machine Placement and Migration for Multi-access Edge Computing

arxiv: v2 [cs.ni] 23 May 2016

CS 655 Advanced Topics in Distributed Systems

WaterSpatial. Harmonic Mean. Execution

CPET 565/CPET 499 Mobile Computing Systems. Lecture 8. Data Dissemination and Management. 2 of 3

RIAL: Resource Intensity Aware Load Balancing in Clouds

Energy-Efficient Load Balancing in Cloud: A Survey on Green Cloud

Approaches for Resilience Against Cascading Failures in Cloud Datacenters

A Mathematical Computational Design of Resource-Saving File Management Scheme for Online Video Provisioning on Content Delivery Networks

A Study of the Performance Tradeoffs of a Tape Archive

CHAPTER 6 ENERGY AWARE SCHEDULING ALGORITHMS IN CLOUD ENVIRONMENT

DeTail Reducing the Tail of Flow Completion Times in Datacenter Networks. David Zats, Tathagata Das, Prashanth Mohan, Dhruba Borthakur, Randy Katz

Hadoop Virtualization Extensions on VMware vsphere 5 T E C H N I C A L W H I T E P A P E R

RIAL: Resource Intensity Aware Load Balancing in Clouds

Tools for Social Networking Infrastructures

Review Article AN ANALYSIS ON THE PERFORMANCE OF VARIOUS REPLICA ALLOCATION ALGORITHMS IN CLOUD USING MATLAB

Towards Energy Proportionality for Large-Scale Latency-Critical Workloads

Memshare: a Dynamic Multi-tenant Key-value Cache

Resource Allocation Strategies for Multiple Job Classes

Consolidating Complementary VMs with Spatial/Temporal-awareness in Cloud Datacenters

Efficient Point to Multipoint Transfers Across Datacenters

ABSTRACT I. INTRODUCTION

Distributed Scheduling of Recording Tasks with Interconnected Servers

Survey of ETSI NFV standardization documents BY ABHISHEK GUPTA FRIDAY GROUP MEETING FEBRUARY 26, 2016

Application of SDN: Load Balancing & Traffic Engineering

Nested QoS: Providing Flexible Performance in Shared IO Environment

Building Adaptive Performance Models for Dynamic Resource Allocation in Cloud Data Centers

QoS routing in DWDM Optical Packet Networks

Interconnection Networks: Topology. Prof. Natalie Enright Jerger

A Strategy to Manage Cache Coherency in a Distributed Mobile Wireless Environment

Dynamic control and Resource management for Mission Critical Multi-tier Applications in Cloud Data Center

Mohammad Hossein Manshaei 1393

TDDD82 Secure Mobile Systems Lecture 6: Quality of Service

AutoTune: Game-based Adaptive Bitrate Streaming in P2P-Assisted Cloud-Based VoD Systems

Nowadays data-intensive applications play a

Resource analysis of network virtualization through user and network

TCP Incast problem Existing proposals

Cutting the Cord: A Robust Wireless Facilities Network for Data Centers

ADAPTIVE AND DYNAMIC LOAD BALANCING METHODOLOGIES FOR DISTRIBUTED ENVIRONMENT

Radware ADC Solution for the Virtual Data Center Dirk Aertgeerts Country Sales Manager Belgium & Luxembourg

LOAD BALANCING IN CONTENT DISTRIBUTION NETWORKS

DriveScale-DellEMC Reference Architecture

RCD: Rapid Close to Deadline Scheduling for Datacenter Networks

Making Gnutella-like P2P Systems Scalable

Balancing the Migration of Virtual Network Functions with Replications in Data Centers

Transcription:

Deadline Guaranteed Service for Multi- Tenant Cloud Storage Guoxin Liu and Haiying Shen Presenter: Haiying Shen Associate professor *Department of Electrical and Computer Engineering, Clemson University, Clemson, USA 1

Outline Introduction Related work PDG design Evaluation Conclusion 2

Introduction Cloud storage Tenant perspective Save capital investment and management cost Pay-as-you-go Service latency vs. revenue Amazon portal: increasing page presentation by 100ms reduces user satisfaction and degrades sales by 1%. Challenge: Reduce the fat-tail of data access latency 3

Introduction Cloud storage Provider perspective Cost-efficient service Cost saving Resource sharing between tenants Energy saving Workload consolidation Encounter problem Unpredictable performance to serve tenants data requests (e.g. service latency) 4

Introduction Problem harmonization Service level agreements (SLAs) [1] (e.g. 99% requests within 100ms) baked into cloud storage services Challenge How to allocate data: non-trivial [1] C. Wilson, H. Ballani, T. Karagiannis, and A. Rowstron. Better Never than Late: Meeting Deadlines in Datacenter Networks. In Proc. of SIGCOMM, 2011. 5

Introduction Our Approach: PDG: Parallel Deadline Guaranteed scheme Goals: traffic minimization, resource utilization maximization and scheme execution latency minimization Assurance: Tenants SLAs Operation: serving ratios among replica servers and creating data replicas Enhancement: prioritized data reallocation for dynamic request rate variation 6

Outline Introduction Related work PDG design Evaluation Conclusion 7

Related work Deadline-aware networks Bandwidth apportion According to deadline Dataflow schedule Prioritize different dataflows Caching system Cache recent requested data Topology optimization Optimized cloud storage Throughput maximization Data availability insurance Replication strategy to minimize cost Problem None of them achieve multiple goals as PDG in cloud storage 8

Outline Introduction Related work PDG design Evaluation Conclusion 9

PDG design Data allocation problem Heterogeneous environment Different server capacities Different tenant SLAs Variations of request rates Multiple constraints SLA insurance Storage/service capacity limitation Multiple goals Network load, energy consumption and computing time minimization Time complexity NP-hard 10

Data reallocation for deadline guarantee as a nonlinear programming 11

PDG design System assumption Each server = M/M/1 queuing system Request arrival rate follows Poisson process The service time follows an exponential distribution Single queue Based on the model, we can derive the CDF of service time of requests Sn: server n; F() sn : CDF of service time; λ sn : request arrival rate, μ sn : service rate 12

PDG design : probability density function that tenant t k s request targets j servers To guarantee SLA: 13

PDG design System assumption To guarantee SLA λ g sn : maximum arrival rate to Sn; K tk : tenant k s deadline strictness, a variable related to the deadline and allowed percentage of requests beyond deadline System requirement to achieve multiple goals with constraints Each server has a request arrival rate lower than λ sn Consolidate workloads of requests to fewer servers Minimize replications and replicate with proximity-awareness Distributed data allocation scheduling g 14

PDG design Tree-based Parallel Process Unsolved servers Underloaded and overloaded servers Each VN (virtual node) runs PDG Serving ratio reassignment Data replication Report unsolved servers to parents 15

PDG design Serving Ratio Reassignment Loop all replicas in overloaded servers to redirect the serving ratio to replicas in underloaded servers Data Replication Create a new replica in the most overloaded server to the most underloaded servers Reassign serving ratio for this replica Loop until no overloaded servers 16

PDG design Workload consolidation Goal Energy consumption minimization Trigger If total available service rate is larger than the g minimum λ sn Procedure Sort servers in an ascending order of λ sn Deactivate the first server If SLA is guaranteed, deactivate next server Otherwise, termination g 17

PDG design Prioritized data reallocation SLA guarantee under request arrival rate variation Select the most heavily requested data items Broadcast within rack for request ratio reassignment Report unsolved servers to load balancer Load balancer conducts PDG to balance requests over racks 18

Outline Introduction Related work PDG design Evaluation Conclusion 19

Evaluation Experimental settings 3000 data servers [6TB, 12TB, 24TB] storage capacity [80,100] service capacity Fat-tree with three layers 500 tenants [100ms, 200ms] Deadline 5% maximum allowed percentage of requests beyond deadline [100, 900] data partitions with request arrival rate follows distribution in [2] [2] CTH Trace. http://www.cs.sandia.gov/scalable IO/SNL_Trace_Data/, 2009. 20

Evaluation Comparison methods Deadline guarantee periodically Random: randomly place data among servers Pisces[3]: storage capacity aware data first fit Deadline: deadline aware first fit CDG: centralized load balancing of PDG Deadline guarantee dynamically PDG_H: PDG using highest arrival rates for all data PDG_NR: PDG without prioritized data reallocation PDG_R: PDG with prioritized data reallocation [3] D. Shue and M. J. Freedman. Performance Isolation and Fairness for Multi-Tenant Cloud Storage. In Proc. of OSDI, 2012. 21

Evaluation Important metrics Excess latency: avg. extra service latency time beyond the deadline for a request SLA satisfaction level: actual percentage of requests within deadline/required percentage QoS of SLA: the minimum SLA satisfaction level among all tenants SLA guarantee Average excess latency: shortest, best performance in deadline violation case SLA ensured: slightly larger than 100% 22

Evaluation Objective achievement Effectiveness of workload consolidation Energy: maximized energy saving Effectiveness of tree-based parallel process Traffic load: minimized network for data reallocation Bottom up process introduces a proximity-aware replication 23

Evaluation Dynamic SLA guarantee and energy savings Performance of SLA guarantee QoS of SLA: PDG_H and PDG_R both guarantee SLA SLA-aware dynamical request ratio and data reallocation Performance of energy saving Energy savings: PDG_R saves more energy than PDG_H Use more servers when needed 24

Outline Introduction Related work PDG design Evaluation Conclusion 25

Conclusion PDG: parallel deadline guaranteed scheme, which dynamically moves data request load from overloaded servers to underloaded servers to ensure the SLAs Mathematical model to give an upper bound on the request arrival rate of each server to meet the SLAs A load balancing schedule to quickly resolve the overloaded servers based on a tree structure A server deactivation method to minimize energy consumption A prioritized data reallocation to dynamically strengthen SLA Future work Real deployment to examine its real-world performance 26

Thank you! Questions & Comments? Haiying Shen shenh@clemson.edu Electrical and Computer Engineering Clemson University