Switched SAS. Sharable, Scalable SAS Infrastructure. Terry Gibbons, Director Software Planning and Strategy, LSI Corporation.

Similar documents
LSI SAS6160 LSI SAS Switch. Configuration Guide

SAS Technical Update Connectivity Roadmap and MultiLink SAS Initiative Jay Neer Molex Corporation Marty Czekalski Seagate Technology LLC

SAS & SATA Combine to Change the Storage Market. Marty Czekalski, Seagate Technology Terry Gibbons, LSI Corporation

Dell PowerVault MD Family. Modular Storage. The Dell PowerVault MD Storage Family

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

Assessing performance in HP LeftHand SANs

Marty Czekalski President, SCSI Trade Association - Emerging Interface and Architecture Program Manager, Seagate Technology

Dell PowerVault MD Family. Modular storage. The Dell PowerVault MD storage family

SAS & SATA Combine to Change the Storage Market

Dell PowerEdge R720xd with PERC H710P: A Balanced Configuration for Microsoft Exchange 2010 Solutions

Nutanix Tech Note. Virtualizing Microsoft Applications on Web-Scale Infrastructure

IBM System Storage DS5020 Express

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

FLASHARRAY//M Business and IT Transformation in 3U

The Oracle Database Appliance I/O and Performance Architecture

FLASHARRAY//M Smart Storage for Cloud IT

LSI and HGST accelerate database applications with Enterprise RAID and Solid State Drives

Microsoft Exchange Server 2010 workload optimization on the new IBM PureFlex System

HCI: Hyper-Converged Infrastructure

Reference Architecture Microsoft Exchange 2013 on Dell PowerEdge R730xd 2500 Mailboxes

Assessing and Comparing HP Parallel SCSI and HP Small Form Factor Enterprise Hard Disk Drives in Server Environments

JMR ELECTRONICS INC. WHITE PAPER

Microsoft SQL Server in a VMware Environment on Dell PowerEdge R810 Servers and Dell EqualLogic Storage

Dell Reference Configuration for Large Oracle Database Deployments on Dell EqualLogic Storage

IBM Europe Announcement ZG , dated February 13, 2007

EDUCATION Harry Mason Marty Czekalski

Designing and Deploying a Cisco Unified Computing System SAN Using Cisco MDS 9000 Family Switches

THE SUMMARY. CLUSTER SERIES - pg. 3. ULTRA SERIES - pg. 5. EXTREME SERIES - pg. 9

Marty Czekalski President, SCSI Trade Association - Emerging Interface and Architecture Program Manager, Seagate Technology

EMC XTREMCACHE ACCELERATES VIRTUALIZED ORACLE

White Paper. EonStor GS Family Best Practices Guide. Version: 1.1 Updated: Apr., 2018

Accelerate Database Performance and Reduce Response Times in MongoDB Humongous Environments with the LSI Nytro MegaRAID Flash Accelerator Card

Mellanox Virtual Modular Switch

Entry-level Intel RAID RS3 Controller Family

IBM Emulex 16Gb Fibre Channel HBA Evaluation

Database Services at CERN with Oracle 10g RAC and ASM on Commodity HW

Interface Trends for the Enterprise I/O Highway

ORACLE FABRIC MANAGER

EMC Virtual Infrastructure for Microsoft Applications Data Center Solution

Vblock Architecture. Andrew Smallridge DC Technology Solutions Architect

Evaluation Report: Improving SQL Server Database Performance with Dot Hill AssuredSAN 4824 Flash Upgrades

NAS for Server Virtualization Dennis Chapman Senior Technical Director NetApp

Dell Exchange 2013 Reference Architecture for 500 to 20,000 Microsoft Users. 1 Overview. Reliable and affordable storage for your business

Storage Access Network Design Using the Cisco MDS 9124 Multilayer Fabric Switch

IBM System Storage DS4800

Virtualization of the MS Exchange Server Environment

IBM PowerVM. Virtualization without limits. Highlights. IBM Systems and Technology Data Sheet

IBM System Storage DS8800 Overview

Deploy a High-Performance Database Solution: Cisco UCS B420 M4 Blade Server with Fusion iomemory PX600 Using Oracle Database 12c

IBM System Storage DS5020 Express

An Oracle White Paper June Enterprise Database Cloud Deployment with Oracle SuperCluster T5-8

Exam : S Title : Snia Storage Network Management/Administration. Version : Demo

EMC Backup and Recovery for Microsoft Exchange 2007

Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads

How Smarter Systems Deliver Smarter Economics and Optimized Business Continuity

THE OPEN DATA CENTER FABRIC FOR THE CLOUD

Storage Area Networks SAN. Shane Healy

Adaptec Series 7 SAS/SATA RAID Adapters

SoftNAS Cloud Performance Evaluation on AWS

LSI MegaRAID Advanced Software Evaluation Guide V3.0

IBM TotalStorage SAN Switch M12

Optimizing and Managing File Storage in Windows Environments

Introduction to NetApp E-Series E2700 with SANtricity 11.10

Using the ThinkServer SA120 Disk Array

STORAGE CONSOLIDATION WITH IP STORAGE. David Dale, NetApp

LEVERAGING FLASH MEMORY in ENTERPRISE STORAGE

VSTOR Vault Mass Storage at its Best Reliable Mass Storage Solutions Easy to Use, Modular, Scalable, and Affordable

PeerStorage Arrays Unequalled Storage Solutions

HPE Synergy HPE SimpliVity 380

vstart 50 VMware vsphere Solution Specification

STORAGE CONSOLIDATION WITH IP STORAGE. David Dale, NetApp

EMC Business Continuity for Microsoft Applications

Finding the pain. Delivering the solution.

Create a Flexible, Scalable High-Performance Storage Cluster with WekaIO Matrix

Nimble Storage Adaptive Flash

2 to 4 Intel Xeon Processor E v3 Family CPUs. Up to 12 SFF Disk Drives for Appliance Model. Up to 6 TB of Main Memory (with GB LRDIMMs)

Building resiliency with IBM System x and IBM System Storage solutions

White Paper. A System for Archiving, Recovery, and Storage Optimization. Mimosa NearPoint for Microsoft

IBM TotalStorage SAN Switch F08

IBM BladeCenter HS21 blade servers offer performance efficiency with powerful, quad-core processor performance

V.I.B.E. Virtual. Integrated. Blade. Environment. Harveenpal Singh. System-x PLM

The Virtualized Server Environment

Microsoft SQL Server 2012 Fast Track Reference Architecture Using PowerEdge R720 and Compellent SC8000

Nimble Storage vs HPE 3PAR: A Comparison Snapshot

Top 5 Reasons to Consider

Cisco HyperFlex HX220c M4 Node

Storage Solutions for VMware: InfiniBox. White Paper

Demartek September Intel 10GbE Adapter Performance Evaluation for FCoE and iscsi. Introduction. Evaluation Environment. Evaluation Summary

Designing a Reference Architecture for Virtualized Environments Using IBM System Storage N series IBM Redbooks Solution Guide

A Comparative Study of Microsoft Exchange 2010 on Dell PowerEdge R720xd with Exchange 2007 on Dell PowerEdge R510

Storageflex HA3969 High-Density Storage: Key Design Features and Hybrid Connectivity Benefits. White Paper

16GFC Sets The Pace For Storage Networks

access addresses/addressing advantages agents allocation analysis

TITLE. the IT Landscape

White Paper. Low Cost High Availability Clustering for the Enterprise. Jointly published by Winchester Systems Inc. and Red Hat Inc.

DAHA AKILLI BĐR DÜNYA ĐÇĐN BĐLGĐ ALTYAPILARIMIZI DEĞĐŞTĐRECEĞĐZ

Introduction to iscsi

SSD Architecture Considerations for a Spectrum of Enterprise Applications. Alan Fitzgerald, VP and CTO SMART Modular Technologies

Unified Storage and FCoE

UCS M-Series + Citrix XenApp Optimizing high density XenApp deployment at Scale

Transcription:

White Paper A B S T R A C T As application intelligence and new techniques for safeguarding data emerge, the subject of bringing storage closer to the server arises. Direct connection to external storage has unique attributes for providing low latency access, fail-over capability, and favorable management and deployment characteristics. This paper brings to light some of the application advancements and how they align with expanded Serial Attached SCSI (SAS) protocol capabilities. To address emerging applications that provide local/remote data replication, streamlined messaging, and increased transaction capability, a new focus has been placed on Switched SAS. This evolving technology is driving new ways to connect storage by delivering increased cable throughput, simple scalability, and do so within a familiar and tightly coupled storage model. By applying this sharable, scalable infrastructure there is no longer a concern over the familiar refrain, purchase more servers to add more storage. Switched SAS Sharable, Scalable SAS Infrastructure Terry Gibbons, Director Software Planning and Strategy, Corporation Terminology: neither Switched SAS nor are defined by the International Committee for Information Technology Standards (INCITS) T10 Technical Committee. For purposes of this discussion a is a managed set of Devices. Switched SAS is a way to conceptualize Expander functionality where T10(1) describes Expander functionality as a connection router and connection manager. Throughout this paper, various deployment options will be discussed. To understand the deployment models, a review of the underlying technology is required. Figure 1 represents SAS Expander functionality. This is an any-to-any connection matrix noting that one could connect initiator-to-initiator but most likely has no reason to do so. Figure 1. Expander Functionality Connection Manager and Router Connection Manager and Router Any-to-Any Connection Cables connect w/4-phys Disks connect w/1-phy Figure 2 shows an outline of a and how it could be used in a broader topology. Note how SAS is currently deployed with four physical links (Phys) per cable and two cables (ports) per SAS Host Bus Adapter (HBA or initiator ). Figure 2. Model Expander Abstraction Illustrates two initiators to many expander-based storage units Each initiator could be in separate host

Figure 3 conveys the concept of Zoning. This is a method available in s that allows individual devices, or groups of devices, to be hidden from selected host-based initiators. In the illustration, we see many drives isolated in Zone A to Host 1. Through the same expander (part of the JBOD), we see two drives in Zone B that are isolated to Host 2. If additional drives were added via JBOD B then they could be seen by both hosts, zoned individually, or even hidden from both hosts as a reserve pool. Drives directly attached to a host are seen only by that host. Host 1 Host 2 ZONE A ZONE B JBOD A JBOD B Figure 3. SAS Zoning SAS Technology SAS is a point-to-point technology where transactions require a complete connection from a host to a target (storage) device. SAS uses port to refer to a collection of links that usually represent traffic over a cable. Depending on the complexity of the topology, a transaction may require several links to connect to one another before completing a connection to a target device. A unique feature of SAS is the bandwidth build into the cables. With each cable representing four SAS links, this represents 4800Mb/s per port, full-duplex at a 6Gb/s link rate. A 16-port (1U / half-width) can accommodate 768Gb/s bandwidth. Later on will be a review of total bandwidth and theoretical performance in server environments. This level of bandwidth aggregation lends itself to low power consumption. While the total package of processor, protocol controller, and case + materials may be very close across SAS, FC, and GbE technologies, the fact that SAS is intended to support many links in a consolidated package provides attractive watt-per-port numbers usually below 5W. This power profile can support 10m passive copper connections. 2010 product introductions support 25m active copper capability and future roadmaps from the SCSI Trade Association(3) indicate support for 100m optical cables. The actual/estimated power requirements for these technologies have yet to be considered as part of the public domain. SAS zoning is not address based. Again, this follows the point-to-point concept where end-devices (initiator or target) may or may not be allowed to complete a transaction based on the sum of the physical connections that must be created. Emerging Applications Hardware and software applications are growing in their capability to support new features and performance criteria within many traditional computing environments. Tiered Storage (e.g. variable access requirements within databases and search engines), Distributed Applications (e.g. email applications), and Virtual Servers (e.g. server racks/pods for a Mega Datacenter) present a need for low latency and a high transaction rate while not sacrificing high availability, data integrity, or data recovery. For purposes of this discussion, Tiered Storage is the concept of how quickly an application needs to access data. Is the data hot, warm, or cool? In this case, the emerging technology is both hardware and its access parameters set by the application. For hot data, like that associated with database transaction logs, high performance computing, or data mining, an SSD may be used to maximize IOPS especially if the transaction size is in the 2KB to 16KB range. If performance is important but not critical then SAS HDD s may be used at the next level. Finally, data warehouse or documentation may reside on large capacity HDD s such as SATA. Figure 4 gives an example of a Tiered Storage deployment where SSD resides in the server, SAS drives have the most immediate access externally, and SATA drives are cascaded. Switched SAS-Sharable, Scalable SAS Infrastructure 2

Multiple SSD SAS Drives SATA Drives Figure 4. Storage Tiering Distributed Applications certainly require network infrastructure to connect computing resources across a campus or across the globe. There are requirements for fail-over, high availability, and data backup/restore. However, these applications have requirements for a high transaction rate that may be best served by SAS storage as part of a larger infrastructure. Emerging applications provide local and remote replication services for backup and disaster recovery. Examples of these services are Cluster and Standby Continuous Replication(2) (CCR/SCR) where local repositories can be managed as external storage attached to a server (see Figure 5). In this case, a port (or physical connection) on the HBA, Switch, or Disk Drive could fail and a back-up path is available. Yet, this also keeps the passive node storage separate from the active node storage. Active Mailer Server Email Traffic Backup Server Node Private Network Redundant es Server A -Too much work? - Can the VM move? Available Server VM? Duplicate local setup at 2nd site (SCR) A multi-ported RAID system ($$) could resolve the mobility issue Primary and Secondary Storage Figure 5. Local/Continuous Replication Figure 6. Old DAS Model An example of Virtual Server deployment is in the large, homogeneous racks of servers + storage that are built for economy. These racks may be further bundled with other racks to create a pod. While many pods may make up a Mega Datacenter, it s the racks and pods that can benefit from externally connected SAS storage as the most efficient configuration because the storage can be shared amongst the servers. Refer to the contrasting images presented in Figures 6 and 7. Emerging applications in the form of virtual server mobility help keep the OpEx in line with the CapEx by simply shifting workloads to another physical server without reconfiguring hardware. Virtual Server mobility is assisted by the concept of exposing all storage to a mobility manager via a. By placing metadata on the disk drives indicating which virtual server image is stored there, the mobility manager can easily tie a virtual server image, and its applications, to a virtual machine at the physical server level. VM Figure 7. VM Mobility Vs. External Storage Server down? Mobility manager moves VM instance to another server Multiple servers w/ multiple VM s VM Images on Disk Accessible to all servers via Switched SAS-Sharable, Scalable SAS Infrastructure 3

IT Requirements Enterprise Datacenters operate with a core set of philosophies that every solution must address. There is a continual growth in data (see Figure 8) and required services that must be managed within highly scrutinized CapEx and OpEx budgets. Figure 8. Storage Growth (TB) by Workload 14,000,000 12,000,000 10,000,000 8,000,000 6,000,000 4,000,000 2,000,000 0 2009 2010 2011 2012 2013 Business Processing Decision Support Collaborative Application Development IT Infrastructure Web Infrastructure Industrial R&D Other Scalable/Sharable. Simplicity of storage expansion, re-provisioning, or varying deployment models. In the Switched SAS model, one may add storage without disruption. Broadcast events notify members in the topology of a change and the new storage is automatically discovered. The storage may be further segregated (zoned) to a specific server, or set of servers, by out-of-band management utilities. This simple philosophy of conveying presence and taking action applies to re-assigning storage to a new function or deploying different numbers of servers and storage units. Simply plug in the components and the topology is understood by all entities. The SAS standard has proven its flexibility in accommodating legacy SAS technology. Expecting this trend to continue allows for future storage expansion even with new, faster technology, thus preserving current investments. All SAS components have options to work at less than their maximum link rate while supporting legacy protocol. Therefore, a 12Gb/s SAS HBA could be deployed in a new server and connect to a 3, 6, or 12Gb/s. The 12Gb/s could connect to 3, 6, or 12Gb/s JBOD s. The JBOD s could connect to 3, 6, or 12Gb/s storage devices. Ease of scaling can be applied to modest setups or more complicated setups that extend into dozens of servers and hundreds of disk drives. By layering es and cascading storage units, the potential for access to thousands of storage devices exists (see Figure 9). Future SAS roadmaps are driving towards larger Switch port-counts such that many configurations avoid layering the Switches. High Availability. Figure 10 shows a Blade deployment scenario. Each Blade Server has a custom form-factor HBA to fit best within the server blade. Note how multiple es can be used to complete a separate IO path (best accomplished using dual-port SAS drives). If any connection fails, the workload can be re-routed. Also, this example uses RAID controllers in the external storage unit to further secure the data. Switched SAS-Sharable, Scalable SAS Infrastructure 4

...... 8 JBODs X 24 HDD...8 Servers......8 Switches... Lower Cost Blade Server Blade Server Blade Server Blade Servers SAS Controller Card Expander SASx36 Expander SASx36 External Storage Expander SASx36 SAS1068 SAS ROC RAID Controller SAS ROC Expander SASx36 SAS1068 RAID Controller Ctlr Higher Bandwidth Figure 9. 1,536 Drive Configuration Figure 10. Blade Chassis w/ Manageability. The SAS infrastructure provides low-touch management. Along with the automated discovery for adding/removing storage, management tasks such as software/firmware updates and new installations are designed to be low-touch. Well behaved software at the host level informs the operating system that a device is busy when a link is dropped for a short period of time. Firmware should be designed to minimize re-discovery times, or not drop links at all, in case of an upgrade. Installations can be replicated by pre-assigning connection attributes since there is no device or address specific knowledge involved in the connection (e.g. deploy the same zoning configuration for identical server/storage racks). Configuration of zoning (Figure 3), as well as a centralized point of contact to view/manage the topology, is through the. By providing a network interface, TCP/IP, and Telnet services, one can manage the Switch out-of-band. Also, the is a logical focal point for physical storage maintenance. By having the Switch in place, layers of storage units may be reduced or eliminated thus making replacement and servicing easier by not tracing cable hops through a challenging set of connections. Theoretical Performance Performance aspects of a are focused on throughput and latency. IOPS isn t considered in this discussion as the disk drive (or SSD) and HBA have a profound effect on IOPS capability. Latency is Switch/Expander centric while throughput is a function of PCI- Express bandwidth in the host, the SAS link rate, disk drive capability, and the Switch s ability to route the traffic. Switch Bandwidth. The concurrent bits of data that a 16-port is intended to support is as follows: - 12Gb/s (full duplex) x 16 ports x 4 phys per port = 768Gb/s - A nominal SAS configuration puts four SAS physical links (phys) into one port, hence, one cable Using a 16-port as an example, we examine the best scenario where there are eight upstream ports to the host servers and eight downstream ports to JBOD storage. Throughput is discussed in terms of half-duplex, despite the full-duplex capability of SAS, to accommodate practical expectations of how servers and storage integrate within a SAS infrastructure. Throughput per port: - 6Gb/s x 4 phys = 24Gb/s per port - (24Gb/s per port) / (8b/10b encoding) = 2.4MB/s per port - 2.4MB/s per port x 88.33% (to accommodate arbitration delays and additional framing) = 2160MB/s per port Throughput in a practical application: - 2160MB/s per port x 8 port pairs = 17,280MB/s per o Requires one link from the initiator to the, and another from the to the next or Expander (i.e. JBOD), to complete a connection from server to storage Switched SAS-Sharable, Scalable SAS Infrastructure 5

The time delay of completing a physical connection impacts overall latency within an Expander-based SAS topology. Times of 100ns or less are to be expected for each expander in the connection path. Referring back to Figure 4, there are two levels of latency. Noting that each JBOD has one expander, the SAS drives each have two hops, one for the Switch and one for the JBOD. Contrast this to the last JBOD in the stack of SATA drives where there are five hops to get to a disk drive. Fanning out storage at the highest level is a requirement for maximizing performance. Actual Performance Test Setup: - Intel-based white-box servers - Single Intel Xeon 5520 microprocessor per server - SAS9200-8e 6Gb/s SAS HBA Initiator - Astek 6Gb/s SAS JBOD - Seagate ST9146803SS, Firmware v0006 - One SAS6160 6Gb/s (all configurations) - IOmeter 2006 in client/server configuration - IO Queue Depth of 8 - Disk drives limited to range of 1000 LBA thus avoiding Seek operations - See Figure 7 for similar layout Test Cases (Sequential Read and Write 256KB Block Size): 1) 8 Servers w/ 1 HBA (x4 SAS) per server, 8 JBOD, 16 HDD per JBOD 1:1 Server to JBOD ratio best performing solution 2) 4 Servers w/ 1 HBA (x4 SAS) per server, 12 JBOD, 16 HDD per JBOD Less processing capacity, more storage capacity 3) 6 Servers w/ 2 HBA (x4 SAS) per server, 4 JBOD pairs (cascaded), 24 HDD per JBOD pair Note how cascading the JBOD s adds an additional hop w/ a minor performance impact 18,000 16,000 14,000 12,000 10,000 16 HDD/JBOD 1 Initiator/JBOD 8,000 6,000 16 HDD/JBOD 3 JBODs/Initiator 4,000 2,000 24 HDD/JBOD Pair 3 Initiators/JBOD Pairs 0 256K Sequential Read 1000LBAs Queue Depth 8 256K Sequential Write 1000LBAs Queue Depth 8 Switched SAS-Sharable, Scalable SAS Infrastructure 6

Analysis With optimal conditions set, the Read and Write tests measure approximately 16,400MB/s or 95% of the goal. As server-to-storage ratios were changed, the expected throughput was cut in half. This is due to the fact that either the server side or storage side was limited to four ports for each of these experiments. Summary The old adage, I need more storage, I ll buy another server, need not apply any longer. The sharable nature at the core of SAS Expander Devices allows creation of new products such as the that is capable of managing traffic connections in a broad topology. New connector technology and further expansion of Expander connection management capabilities offer a roadmap to greater possibilities in the future. Furthermore, SAS is served by new techniques for data management and replication. These new applications have a two-fold purpose: First, to supply new ways to provide high availability and reliability; Second, to challenge the status quo in terms of new levels of transaction processing and throughput. SAS protocol and the provide functionality for simplifying the storage environment and reducing operating expense by providing: - Simplified cabling that reduces/eliminates hops that add performance delays and physical layout challenges - Equal performance across all drives and servers - Single configuration, reporting, and control point via a - Stateless servers: any-to-any topology, enables virtualization mobility - Access control via SAS Zoning SAS is an excellent infrastructure for making affordable, manageable, and compact storage deployments an attractive alternative to network based storage designs. List of references (1) Working Draft Project American National T10/2124-D Standard: Information Technology SAS Protocol Layer (SPL) (2) Concept from Microsoft Exchange Server 2007 at technet.microsoft.com (3)SCSI Trade Association (www.scsita.org): Advanced Connectivity Solutions Unleash SAS Potential SCSI Trade Association White Paper (Author: Harry Mason, Contributor: Jay Neer) Switched SAS-Sharable, Scalable SAS Infrastructure 7

For more information and sales office locations, please visit the web sites at: lsi.com lsi.com/channel lsi.com/contacts, and Design logo, are trademarks or registered trademarks of Corporation. All other brand and product names may be trademarks of their respective companies. Corporation reserves the right to make changes to any products and services herein at any time without notice. does not assume any responsibility or liability arising out of the application or use of any product or service described herein, except as expressly agreed to in writing by ; nor does the purchase, lease, or use of a product or service from convey a license under any patent rights, copyrights, trademark rights, or any other of the intellectual property rights of or of third parties. Copyright 2010 by Corporation. All rights reserved. October 2010