Technologies of ETERNUS6000 and ETERNUS3000 Mission-Critical Disk Arrays

Similar documents
ETERNUS DX60 and DX80

Fujitsu s Chipset Development for High-Performance, High-Reliability Mission- Critical IA Servers PRIMEQUEST

White Paper Features and Benefits of Fujitsu All-Flash Arrays for Virtualization and Consolidation ETERNUS AF S2 series

FUJITSU ETERNUS DX8000 SERIES DISK STORAGE SYSTEMS FOR LARGE ENTERPRISES

Slide 0 Welcome to this Web Based Training session introducing the ETERNUS DX80 S2, DX90 S2, DX410 S2 and DX440 S2 storage systems from Fujitsu.

The term "physical drive" refers to a single hard disk module. Figure 1. Physical Drive


IBM System Storage DS5020 Express

IPCOM EX Series for Realizing Network Stability and Safety

Innovative Technologies Realizing Compact PRIMERGY TX120 Server

Application-Oriented Storage Resource Management

VX3000-E Unified Network Storage

FUJITSU Storage ETERNUS AF series and ETERNUS DX S4/S3 series Non-Stop Storage Reference Architecture Configuration Guide

NAS System. User s Manual. Revision 1.0

Dell Fluid Data solutions. Powerful self-optimized enterprise storage. Dell Compellent Storage Center: Designed for business results

Dell PowerVault MD Family. Modular storage. The Dell PowerVault MD storage family

SYSTEM UPGRADE, INC Making Good Computers Better. System Upgrade Teaches RAID

QuickSpecs. Compaq Smart Array 431 Controller M ODELS

EMC Symmetrix DMX Series The High End Platform. Tom Gorodecki EMC

Always On White Paper for Fujitsu ETERNUS Storage System

Introduction to NetApp E-Series E2700 with SANtricity 11.10

Dell PowerEdge R720xd with PERC H710P: A Balanced Configuration for Microsoft Exchange 2010 Solutions

The benefits of. Clustered storage offers advantages in both performance and scalability, but users need to evaluate three different architectures.

Definition of RAID Levels

VX1800 Series Unified Network Storage

Analysis of Fujitsu ETERNUS4000 Models 80, 100, 300, and 500, Feature/Function Analyst: Tom Trainer

TECHNICAL SPECIFICATIONS

SPC BENCHMARK 2 EXECUTIVE SUMMARY FUJITSU LIMITED FUJITSU STORAGE SYSTEMS ETERNUS DX8700 S2 SPC-2 V1.4

QuickSpecs. Models. HP Smart Array 642 Controller. Overview. Retired

OceanStor 5300F&5500F& 5600F&5800F V5 All-Flash Storage Systems

Storage Area Networks: Performance and Security

Dell EMC SCv3020 7,000 Mailbox Exchange 2016 Resiliency Storage Solution using 7.2K drives

VSTOR Vault Mass Storage at its Best Reliable Mass Storage Solutions Easy to Use, Modular, Scalable, and Affordable

Disk Storage Systems. Module 2.5. Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved. Disk Storage Systems - 1

IBM System Storage DS6800

Data Sheet Fujitsu ETERNUS DX200 S3 Disk Storage System

EMC CLARiiON CX3-40. Reference Architecture. Enterprise Solutions for Microsoft Exchange 2007

EMC Business Continuity for Microsoft Applications

White Paper FUJITSU Storage ETERNUS DX S4/S3 series Extreme Cache/Extreme Cache Pool best fit for fast processing of vast amount of data

S SNIA Storage Networking Management & Administration

The GP7000F UNIX servers

CONFIGURING ftscalable STORAGE ARRAYS ON OpenVOS SYSTEMS

Technical Note P/N REV A01 March 29, 2007

Data Sheet Fujitsu ETERNUS DX500 S3 Disk Storage System

Data Sheet FUJITSU Storage ETERNUS DX100 S3 Disk System

SOFTWARE GUIDE VT-100. imageraid Series Storage System. (Text-based User Interface)

Sun Fire V880 System Architecture. Sun Microsystems Product & Technology Group SE

RAID FIBRE TO S-ATA/SAS

Models Smart Array 6402/128 Controller B21 Smart Array 6404/256 Controller B21

Take control of storage performance

EMC CLARiiON CX3 Series FCP

EMC VMAX 400K SPC-2 Proven Performance. Silverton Consulting, Inc. StorInt Briefing

IBM System Storage DS5020 Express

Cisco UCS C24 M3 Server

IBM Netfinity ServeRAID-3H and -3L Ultra2 SCSI Adapters

IBM. IBM PC Servers. IBM SSA RAID Cluster Adapter. Hardware Maintenance Manual Supplement. October 1997

Technical White Paper FUJITSU Storage ETERNUS AF and ETERNUS DX Feature Set

White Paper. EonStor GS Family Best Practices Guide. Version: 1.1 Updated: Apr., 2018

TECHNICAL SPECIFICATIONS + TECHNICAL OFFER

Configuring a Single Oracle ZFS Storage Appliance into an InfiniBand Fabric with Multiple Oracle Exadata Machines

Chapter 10: Mass-Storage Systems

RocketU 1144CM Host Controller

New HPE 3PAR StoreServ 8000 and series Optimized for Flash

Vendor: EMC. Exam Code: E Exam Name: Cloud Infrastructure and Services Exam. Version: Demo

IBM xseries 460 servers XpandOnDemand with eserver x3 architecture

IBM System x3950 servers feature dual-core Intel Xeon MP processors

RocketU 1144BM Host Controller

RAID EzAssist Configuration Utility User Reference Guide

Demartek December 2007

Data Sheet FUJITSU Storage ETERNUS DX600 S3 Disk System

Chapter 10: Mass-Storage Systems. Operating System Concepts 9 th Edition

USER S GUIDE. Integrated RAID for SAS. April Version 1.1 DB

DELL Reference Configuration Microsoft SQL Server 2008 Fast Track Data Warehouse

FUJITSU SPARC ENTERPRISE SERVER FAMILY MID-RANGE AND HIGH- END MODELS ARCHITECTURE FLEXIBLE, MAINFRAME- CLASS COMPUTE POWER

Monitoring ETERNUS DX systems with ServerView Operations Manager

DATA DOMAIN INVULNERABILITY ARCHITECTURE: ENHANCING DATA INTEGRITY AND RECOVERABILITY

SAN Storage Array Workbook September 11, 2012

IBM System Storage DS4800

REC (Remote Equivalent Copy) ETERNUS DX Advanced Copy Functions

LEVERAGING A PERSISTENT HARDWARE ARCHITECTURE

IBM IBM Open Systems Storage Solutions Version 4. Download Full Version :

Data Sheet FUJITSU Storage ETERNUS DX200F All-Flash-Array

Cluster Configuration Design Guide (Linux/PRIMECLUSTER)

How to Speed up Database Applications with a Purpose-Built SSD Storage Solution

Lenovo RAID Introduction Reference Information

VERITAS Storage Foundation 4.0 TM for Databases

FOUR WAYS TO LOWER THE COST OF REPLICATION

Dell PowerVault MD Family. Modular storage. The Dell PowerVault MD storage family

Vendor must indicate at what level its proposed solution will meet the College s requirements as delineated in the referenced sections of the RFP:

Oracle Database ILM Solution based on Fujitsu ETERNUS DX/SPARC Enterprise

White paper PRIMEQUEST 1000 series high availability realized by Fujitsu s quality assurance

DELL EMC UNITY: HIGH AVAILABILITY

JMR ELECTRONICS INC. WHITE PAPER

THE SUMMARY. CLUSTER SERIES - pg. 3. ULTRA SERIES - pg. 5. EXTREME SERIES - pg. 9

The Vess R2000 Series

EonStor DS - SAS Series High-Availability Solutions Delivering Excellent Storage Performance

Data Sheet Fujitsu M10-4S Server

IBM TotalStorage Enterprise Storage Server Model 800

Hitachi Virtual Storage Platform Family

Apace Systems. Avid Unity Media Offload Solution KIT

Transcription:

Technologies of ETERNUS6000 and ETERNUS3000 Mission-Critical Disk Arrays V Yoshinori Terao (Manuscript received December 12, 2005) Fujitsu has developed the ETERNUS6000 and ETERNUS3000 disk arrays for mid-range and high-end IT market segments. These segments have been growing because IT systems have shifted from server-centric processing architectures to information-centric architectures. This paper introduces the system architecture; highreliability, high-performance, and high-availability features; and technologies of the. Section 2 describes the controller modules, channel adapters, device adapters, routers, and device enclosure features. Section 3 describes the high-reliability features such as cyclic mirroring, block check codes, DB Guard, and redundant copying. Section 4 describes the high-availability RAID migration and logical device expansion features. Lastly, Section 5 describes the high-performance technologies of these RAID subsystems. 1. Introduction Fujitsu introduced its first ETRNUS6000 and ETERNUS3000 disk arrays in 2003 and then improved their performance in 2004 and 2005. Recently, IT systems have been shifting paradigm from architectures focused on server-centric processing to information-centric architectures. This shift is making storage systems even more important in the storage, distribution, and utilization of information. The result is an ongoing process of storage system integration and networking that will continue into the foreseeable future. To meet the growing requirements for IT systems, Fujitsu provides the ETERNUS6000 storage systems for high-end, multi-platform environments and the ETERNUS3000 storage systems for mid-range IT systems. Both subsystems are designed to form part of an advanced mission-critical IT system. 2. System architectures Figures 1 and 2 show the architectures of the ETRNUS6000 and ETRNUS3000. The consist of a controller enclosure (CE) and two or more drive enclosures (s). The ETERNUS3000 can have up to 16 s, and the ETERNUS6000 can have up to 68. 2.1 Controller enclosure 2.1.1 Controller modules of The controller modules (s) are the main components of the controller enclosure. Each has a processor and cache and controls the entire storage system. To achieve a suitable performance and reliability for high-end RAID systems, the ETERNUS6000 can have up to four s. 2.1.2 Channel adapters of The channel adapters (s) provide connections between the and servers. The ETERNUS6000 supports a wide 2 FUJITSU Sci. Tech. J., 42,1,p.2-8(January 2006)

To drive enclosures To servers To drive enclosures Data PNB 8 paths max. Maximum speed 1.05 GB/s PNB To servers Figure 1 ETERNUS6000 architecture. variety of host interfaces. Its s are available in five types: 4 Gb/s Fibre Channel, 1 Gb/s iscsi, FCLINK, OCLINK, and BMC. The types can be freely combined, and up to 16 s can be installed. When 4-port Fibre Channel s are used, up to 64 ports are available for connection to the host. Each of the ETERNUS6000 has a microprocessor to distribute the I/O workload from servers between the s and s. The ETRNUS3000 supports 4 Gb/s Fibre Channel and 1 Gb/s iscsi s. These s have no microprocessors and are controlled directly by the s. 2.1.3 Device adapters of The device adapters (s) are modules that provide connections between the s and hard disk drives (HDDs). Each has a 2 Gb/s Fibre Channel interface. Each of the ETERNUS6000 has a microprocessor to distribute the many HDD operations between the s and s. The ETERNUS6000 can have up to eight s. Each of the ETERNUS3000 has a controller LSI. These LSIs have no microprocessor and are controlled directly by the s. PNB: PCI Node Link Bridge Figure 2 ETERNUS3000 architecture. 2.1.4 Routers of ETERNUS6000 The routers (s) of the ETERNUS6000 organically interconnect the s, s, and s and provide high-speed communication between them. To achieve system redundancy, each is connected to each. Up to four routers can be installed. 2.2 Drive enclosures of The s can hold up to 15 HDDs plus two backend FC connectivity cards. They also have dual power supply units and fans. 3. High-reliability features 3.1 Component redundancy and hot swap of The most important requirement in integrated storage systems is robust reliability that ensures the safe use of stored. The meet these requirements with full-component redundancy, hot-swappable hardware/firmware, and a wide variety of highreliability and high-availability features. All major components included in the controller and drive enclosures are redundant as standard to ensure continuous operation even when one of them fails. Moreover, these major components are all hot-swappable. This enables on-line recovery from component failures to ensure ongoing redundancy and system continuity. FUJITSU Sci. Tech. J., 42,1,(January 2006) 3

3.2 Cyclic mirroring of ETERNUS6000 s The ETERNUS6000 uses a high-reliability controller system called Cyclic Mirroring, which becomes operational when three or more s are installed. This system minimizes the impact on performance and redundancy when a or a s cache fails. When four s are installed as shown in Figure 3, cache write is mirrored on two s in case one of them fails. If one of the s fails, the Cyclic Controller Module system takes it offline and reconfigures the remaining three so that all write is mirrored among them. Data is assigned cyclically across the three s in a round-robin approach that does not rely on fixed pairs. This means that if a fails, the overall system performance is only degraded by one-quarter or less. one are connected to four different drive enclosures. The ETERNUS6000 forms a RAID-5 group using four selected HDDs from the four drive enclosures, which means the RAID-5 configuration is always fixed at 3+1. However, the ETERNUS6000 can keep the RAID-5 group accessible even when an entire drive enclosure becomes unusable, while limiting the impact of the failure to a single HDD. This enables the RAID s restoration feature to be used. 3.4 Block check code of When receiving from a server, the generate a proprietary block checking code (BCC) from calculations made on the block. Figure 5 shows the BCC process of the ETERNUS6000. In this process, the BCC 3.3 Enhanced reliability RAID-5 of ETERNUS6000 When a failure occurs in a conventional RAID-5 configuration consisting of HDDs in a drive enclosure, the entire drive enclosure becomes unusable and the cannot be accessed. However, in the ETERNUS6000, the HDDs of a RAID-5 group are located in different drive enclosures to ensure robust RAID-5 reliability. As shown in Figure 4, four interfaces belonging to -1-2 -3-4 Degradation -1-2 -3-4 : Drive enclosure Figure 3 Cyclically mirrored s. Figure 4 RAID-5 with enhanced reliability. 4 FUJITSU Sci. Tech. J., 42,1,(January 2006)

Read Server Generate Database ETERNUS6000 Read Generate ETERNUS6000 and cache and cache DISK Figure 5 Block check code feature of ETERNUS6000. DISK component is recorded together with the in cache memory and on the HDDs. For writing, the BCC is used for integrity verification when is sent from the controller to the HDDs. For reading, the BCC is used to verify integrity when is sent from HDDs to the controller and when is sent from the controller to a server. Although the BCC does not provide a correction feature such as ECC, BCC can verify the location and order of ; for example, it can verify whether stripped across multiple HDDs has been rebuilt in the correct order. 3.5 DB Data Guard of ETERNUS6000 BCC is not perfect because it does not cover corruption that occurs between the storage system and server. This is because the ETERNUS6000 will only generate BCCs based on the corrupted. This is a critical issue, especially for bases because corruption can seriously impact the integrity of a base. To solve this issue, we developed DB Data Guard, which is a assurance feature that combines the storage system s hardware and base software. When writing to a storage system, the base generates check codes and adds them to the (Figure 6). The storage system and the base share information, including the check Figure 6 DB Data Guard. code generation logic and the positions where the check codes are added. The ETERNUS6000 has a special ASIC in its s that calculates and verifies this check code. If a detects that received from the server is corrupted, it reports the detection and discards the so it does not overwrite any in the s cache. This prevents the use of corrupted and protects the integrity of the base. 3.6 Redundant Copy feature of The ETRNUS6000 and ETERNUS3000 constantly monitor all HDDs. If an HDD generates errors at more than a specified rate, the controller judges that the HDD is about to fail. In this case, the HDD is swapped using the Redundant Copy feature (Figure 7). This feature maintains redundancy even during HDD replacement and minimizes the restoration time. In RAID configurations, when an HDD fails, is restored from the other HDDs to prevent loss. Although redundancy cannot be maintained during restoration (even though it only takes a short time), the Redundant Copy feature does not isolate the HDD in pre-failure status until it FUJITSU Sci. Tech. J., 42,1,(January 2006) 5

RAID-5 RAID5 (3+1) LUN0 LUN1 Pre-failure detection Hot spare Concurrently migratable RAID-5 RAID0+1 (4+4) LUN0 LUN1 LUN2 LUN: Logical Unit Number Copied without losing redundancy Figure 8 RAID migration. RAID-5 Isolated after rebuilding in RAID-5 Figure 7 Redundant Copy. has been copied to a hot-spare HDD. This ensures redundancy even during HDD replacement. If an HDD fails in a conventional RAID-5 configuration, its is rebuilt from the other HDDs before it is copied to a hot-spare HDD. However, Redundant Copy just copies the pre-failure HDD directly to a hot-spare HDD, which reduces the time needed to replace an HDD and the associated system load. 4. High-availability features The include a wide variety of new features that enable flexible system reconfiguration and capacity expansion. Conventionally, a capacity expansion of a RAID group requires a cumbersome procedure: the operator must save the, add a new disk to the RAID group, redefine the RAID configuration, and finally rewrite the back to the revised disk configuration. Moreover, the saved cannot be accessed during this procedure. 4.1 RAID migration of ETENUS6000/3000 In some systems, RAID migration may need to be done to a differently configured RAID (Figure 8). For example, this may be necessary for migration from a RAID-5 group to a RAID-1 group or from a 10 000 rpm HDD RAID group to a 15 000 rpm HDD RAID group in order to improve performance. Conventionally, such operations interrupt system operation and require a cumbersome procedure including backup, configuration of the new RAID area, and restoration to the new RAID group. The ETERNUS6000 provides a RAID Migration feature that performs all of these procedures without interrupting system operation. Moreover if the destination RAID has free space, it can concurrently define a new Logical Unit Number (LUN). 4.2 Logical device expansion of ETERNUS3000 Logical device expansion is a function that automatically performs the complex processing required to expand the RAID group (Figure 9). By using this function, the existing is automatically relocated, and a new LUN can be defined using the newly generated space. The figure shows an example of adding a new 73 GB disk to a RAID5 (3+1) group to make a new RAID5 (4+1) group. 6 FUJITSU Sci. Tech. J., 42,1,(January 2006)

RAID5 (3+1) New disk 73 GB 73 GB 73 GB 73 GB 73 GB RAID5 (4+1) LUN0 146 GB LUN1 73 GB 73 GB 73 GB 73 GB 73 GB 73 GB Figure 9 Logical device expansion. LUN0 146 GB LUN1 73 GB LUN2 73 GB 5. High-performance features The achieve high performance in their RAID subsystems because their s employ high-speed CPUs. Also, the s are connected to fast Interconnect technology and fast HDDs to achieve high-speed movement. 5.1 ETERNUS6000 technology 5.1.1 Processors The ETERNUS6000 uses high-performance processors in each of its s, s, and s and supports a maximum cache of 32 GB. The s, for example, have leading-edge 3.2 GHz processors. The s and s also have high-performance RISC processors to control the interface hardware (i.e., the Fibre Channel controller). 5.1.2 High speed, multi-channel interconnection The ETERNUS6000 features ASIC s that were developed using Fujitsu s original technology. The s provide faster interconnects between the s and s/s. Each has a total bandwidth of 4.2 GB/s. To improve duplication from the s to s, the s duplicate the they receive and send it to two s at the same time ( forking). Generally, in RAID subsystems, the write throughput performance is less than that for read operations. The forking feature prevents a performance reduction during write operations. The ETERNUS6000 can simultaneously transfer up to 96 blocks between the adapters and the cache, which guarantees highly autonomous operation of the system s processors. 5.2 ETERNUS3000 technology 5.2.1 Processors The ETERNUS3000 provides unparalleled performance by using cache sizes of up to 16 GB and using 1.26, 2.8, or 3.2 GHz processors. 5.2.2 DTC In addition to the basic hardware advantages described above, the ETERNUS3000 achieves even faster and more stable performance by using innovative Fujitsu technologies such as Fujitsu s unique Dynamic Disk Traffic Control (DTC) feature (Figure 10). This feature optimizes HDD access on ETERNUS3000 storage systems based on the method of access being used at the time. This ensures a quick response to individual server requests and minimum processing times for background serial access operations such as backup and recovery. When both sequential and random accesses occur simultaneously, for example, when is being restored while random requests are still being processed from the server, they are separated into optimum extents and processed alternately on a time interval basis. This ensures that the optimum access methods are used, even during high-traffic periods. 5.2.3 interconnection The ETERNUS3000 also makes use of a Fujitsu technology innovation called the PCI Node Link Bridge (PNB) (Figure 2), which provides FUJITSU Sci. Tech. J., 42,1,(January 2006) 7

Controller Random access DTC Sequential access Random access Time sharing Disk Sequential access Time Random access and sequential access are processed alternately. Random access only Sequential access only Slot (= fixed time) Slot Slot Slot Figure 10 Dynamic Disk Traffic Control (DTC). faster inter-cache duplication, resulting in improved update processing. When a controller receives a write request from a server, the controller uses PNB hardware to duplicate the in another controller without access interruption. This method reduces the overhead that is usually caused by inter-controller communication for such duplication. PNB supports transfer rates of up to 1.05 GB/s and up to 8 intercontroller communication paths for ultra-fast cache duplication. 6. Conclusion This paper described Fujitsu s ETERNUS6000 and ETERNUS3000 disk arrays, which achieve high performance and high availability while giving priority to reliability. These two subsystems use common technologies such as hot swapping, BCC, and redundant copying to improve their reliability. The ETERNUS6000 uses s, s with processors, DB Data Guard, and RAID migration to provide the performance and reliability required in high-end products. The ETERNUS3000 achieves high performance by using DTC and PNB interconnection. In the future, customers will request RAID subsystems with a wider variety of configurations to meet their individual IT needs. Therefore, a theme for future development at Fujitsu will be the fusion of mid-range and high-end products while maintaining scalability. Yoshinori Terao received the B.S. degree in Applied Physics from Okayama University of Science, Okayama, Japan in 1989. He joined Fujitsu Ltd., Kawasaki, Japan in 1989, where he has been engaged in development of storage system products. E-mail: terao.yoshinori@jp.fujitsu.com 8 FUJITSU Sci. Tech. J., 42,1,(January 2006)