PowerHA SystemMirror 7.2. Split Brain Handling through SCSI PR Disk Fencing
|
|
- Katherine Barrett
- 6 years ago
- Views:
Transcription
1 PowerHA SystemMirror 7.2 Split Brain Handling through SCSI PR Disk Fencing Authors: Abhimanyu Kumar Prabhanjan Gururaj Rajeev S Nimmagada Ravi Shankar Page 1 of 23
2 Table of Contents...1 1Introduction Cluster Split Brain Condition Disk Fencing Disk Fencing Pre requisites: SCSI-3 Persistent Reserve capabilities support EMC Storage and SCSI Hitachi Storage & SCSI Setup Prerequisites SMIT Panel clmgr Command Runtime Considerations References...22 Page 2 of 23
3 1 Introduction This blog introduces a new feature Disk Fencing Quarantine Policy in the PowerHA SystemMirror 7.2 release. This feature will provide protection against rare Cluster Split brain conditions in a PowerHA SystemMirror Cluster. Page 3 of 23
4 2 Cluster Split Brain Condition Cluster based High Availability solutions depend on redundant communication channels between the nodes in a cluster enabling health monitoring of those nodes. This communication is a critical function that enables the cluster to start workload on an alternate node in the cluster when the production node goes down.. Figure 1. below shows a typical cluster deployment used to provide a High Availability (HA)_topology. This cluster has 3 network connections for redundancy and also a disk which is used for heartbeat purposes. This example configuration provides 4 channels of redundant communication between the nodes in the cluster. Note that in this case, the application is operating from an active LPAR on System 1 and the system 2 LPAR is in passive/standby mode ready to take over the workload in case the active or primary LPAR fails. Each node tracks the health of its partner by monitoring the heartbeats and other communications. For the heartbeats to be exchanged, there has to be a good communication channel between the nodes. Hence it is essential to have as many redundant communication channels as possible between the nodes to avoid any false failures. Fig 1: Cluster High Availability: Redundant communication channels Page 4 of 23
5 However, there could be scenarios where all the communication channels between the nodes are broken. Examples of such scenarios include: 1. In extremely rare occasions due to hardware errors (sick but not dead type of errors eg: all the IO fabrics had a freeze for one ndoe) and such it may be possible that primary LPAR freezes for long periods of time resulting in no IO communication to occur from the LPAR. In this case Standby LPAR does not receive any heartbeats from Active LPAR for a pre-determined duration (Node Failure Detection Time) and then declares Active LPAR to be dead. However after that declaration it is possible that Active LPAR unfreezes continuing the IO. This scenario would result in a Cluster split for a duration of time. This could have major impacts to data integrity as explained later. 2. Some IO failures could potentially result in an extended blackout window for IO. For example if the PCI bus master failed and therefore froze the entire PCI infrastructure for a duration of time resulting no IO activity for the period If this IO blackout time is more than the Node Failure detection time threshold, then standby LPAR would declare the primary to have failed (false failure). This would result in cluster enter split state temporarily during the IO black out time. Cluster split results in two sides. In the example of 2 node cluster, each side would consist of one node. Note that if more nodes are part of the cluster, then each side could have more nodes than 1. For example in a 4 node cluster, split could occur such that sides could be: 1 node and 3 nodes 2 nodes and 2 nodes These sides are also called as partitions or islands (of nodes). A Cluster split could result in Application being started on the Standby LPAR incorrectly due to the false node failure detection. This is shown in the figure below: Page 5 of 23
6 Fig 2: Cluster Split Condition: Incorrect and Duplicate application start. As can be seen Standby incorrectly declares primary LPAR to have failed and starts the application resulting in the application being active on both LPARs simultaneously. This would have the disastrous result of both application writing to the shared disks resulting in data corruption. Cluster splits are rare but unavoidable in extreme cases such as sick but not dead components in the environment. It is critical to protect the cluster environment against these rare conditions. PowerHA SystemMirror v7 has supported many capabilities such as disk tie breaker to protect against the Cluster split conditions. PowerHA SystemMirror Version 7.2 introduces new capabilities to protect against the split brain condition. Two quarantine policies are introduced to handle split scenarios: 1. Active Node Halt Policy (ANHP) 2. Disk Fencing In this blog we will review the Disk Fencing in some detail. Page 6 of 23
7 3 Disk Fencing Disk Fencing policy will fence out the disks of all the volume groups added to any resource group using SCSI Persistent Reserve Protocols. This will ensure that only one island will have access to the disks and data will remain protected. This flow ensures that Application will be operating from only one node in the cluster at any time. PowerHA SystemMirror registers at the disks of all the volume groups which are a part of any resource group. Fig 3: Cluster Split Condition: Disk Fencing flow As shown in the Figure 3, Standby LPAR reaches out to the storage and then requests that the Active LPAR disk access be revoked (pre-empted). Storage will then block any write accesses from the Active LPAR (even if it returns from sick to healthy state). Standby LPAR will bring up Application/RG only if it is able to successfully fence out the Active LPAR in regards to the RG related disks. If Standby encounters any errors while fencing out the Active LPAR, then workload will not be brought up and administrator would need to review and correct the environment as necessary and then bring up the RG manually if necessary. Some of the key attributes of PowerHA SystemMirror Disk fencing are: 1. Disk fencing applies to Active-Passive Cluster deployment. Disk fencing is not suitable for RGs of type online on all nodes deployment and is not supported. Page 7 of 23
8 2. Disk Fencing is supported for the entire cluster. So it can be enabled or disabled at the cluster level 3. PowerHA SystemMirror does the key setup and management related to SCSI-3 reserves. Administrator is not expected to do any SCSI-3 key management 4. All the disks managed as part of Volume groups of various Resource Groups (RG) are managed for disk fencing 5. Disk fencing can be used in the mutual takeover configurations. If multiple RGs exist in the cluster, administrator needs to choose one RG to be the most critical RG. This RG s relative location at the time of split decides which side will win after a split (it would be the side where the critical RG was not running before split This side is considered the standby side for the critical RG at that time) PowerHA SystemMirror 7.2 uses as much information as possible from the cluster to determine the health of the partner nodes. For example, if the Active LPAR were going to crash, it will try to send a Last Gasp message to the standby LPAR before terminating. These types of notifications help to insure that the standby LPAR to be certain of the death of the Active LPAR and hence can take workload ownership safely. However, there are cases where the standby LPAR is aware that the active LPAR is not sending heartbeats but is not sure of the actual status of the active LPAR. In these cases, the standby LPAR will declare that the active LPAR has failed after waiting for time duration of Node Failure Detection Time. At that time, since the standby partition is not sure of the health of the active LPAR, it will fence out the all the disks before bringing the resource groups online. If it failed to fence even a single disk of any volume group, resource group will not be brought online. 3.1 Disk Fencing Pre requisites: 1. Storage systems should be enabled for SCSI-3 PR capabilities for all the disks managed as part of the disk fencing 2. All the disks to be managed for Disk fencing should not be in use when the disk fencing is enabled (that is all the VGs should be offline) 3. Disks should be free of any reserves before starting PowerHA SystemMirror configuration. Tools are provided to release any reserves SCSI-3 Persistent Reserve capabilities support One of the key requirements for Disk Fencing is that all the disk being used should support SCSI 3 Persistent Reserve (PR) Protocols. Note that some of the storage subsystems do not enable the PR capabilities by default. These storage sub systems provide commands or graphical interfaces to enable the PR capabilities. Some storage specific guidelines are provided below. Note that these instructions might be out of date depending on the storage model etc. Please refer to the Storage vendor documentation for exact methods to enable the SCSI-3 capabilities. Page 8 of 23
9 EMC Storage and SCSI-3 EMC disks do not support SCSI-3 capabilities by default. If you try to configure Disk fencing without enabling the capability in EMC storage, you will get an error. Enable SCSI-3 reservation capabilities in EMC storages (VMAX and DMX) by enabling SPC2 and SC3 capabilities for each disk assigned to the Volume groups to be managed by PowerHA SystemMirror. Please refer to the EMC documentation for detailed instructions to enable the SCSI-3 capability. Following is an example set of steps tested with EMC storage VMAX (note that many of these commands are part of the EMC software packages installed on PowerHA/AIX LPAR): For each disk, do the following (note that while doing these operations disks should not be in use (all the VGs with these disks should not be vary d on etc). Ideally make sure that disks are not in use on any of the nodes in the cluster.): 1. Find the device/disk id in the EMC storage subsystem 2. Enable the SCSI-3 PR capabilities in the EMC storage subsystem 3. Rediscover the disk/s fresh in AIX: a. Remove the device/disk b. Run cfgmgr to discover the disk/s 4. Verify that the SCSI-3 capabilities are enabled for the disk/s Once these steps have been completed, configure PowerHA SystemMirror disk fencing. Here are the example commands: Retrieve the Disk identity from EMC storage Symmetrix ID(sid) Logical device(device id) # powermt display dev=hdiskpowerx Pseudo name=hdiskpowerx Symmetrix ID= Logical device ID=0036 Device WWN= xxxx state=alive; policy=symmopt; queued-ios=0 Enable SCSI-3 capability using the disk identitity Rediscover Disk in AIX Verify the SCSI-3 # symconfigure -sid cmd "set device 0036 attribute=scsi3_persist_reserv;" commit -v -noprompt # rmdev -Rdl hdiskpowerx # cfgmgr #/usr/symcli/bin/symdev -sid show 0036 grep Page 9 of 23
10 capability is enabled in Storage SCSI SCSI-3 Persistent Reserve: Enabled Hitachi Storage & SCSI-3 Hitachi disables SCSI-3 capability by default. You need to manually enable these capabilities for the disk groups assigned to the PowerHA SystemMirror LPARs for shared VG management. Enable HMO 2 and HMO 72 options through the Hitachi provided graphical management interface software for the storage. Note that the graphical interface picture is from Hitachi and please refer to Hitachi documentation for more details. Page 10 of 23
11 3.2 Setup This section explains how to setup PowerHA SystemMirror 7.2 to enable Disk Fencing. Before enabling the disk fencing mechanism one must ensure that all the shared disks used in the cluster are capable of SCSI3 protocols. Details are provided in the prerequisites (section 3.2.1) Setup for this policy could be done in one of two ways: 1. Using SMIT panels (Section 3.2.2) 2. Using clmgr command line (section 3.2.3) Page 11 of 23
12 3.2.1 Prerequisites Currently SCSI3 protocols are not supported on ISCSI disks. For EMC disks a minimum version of PowerPath v6.0.1 is needed. Set Flags SCSI3_persist_reserv, SPC2, SC3 For Hitachi disks HMO 2 and HMO 72 should be set Minimum code to support HMO72 is /00 A resource group must be defined as a critical RG in the cluster. This Critical RG has to span on all nodes in the cluster. Critical RG cannot have start up policy as Online on All Available nodes. Critical RG cannot be a child as part of any relationship. Parent-child or Start-after Critical RG should have higher priority in Location dependencies. If Active Node Halt policy is also enabled on the cluster, only one RG will be chosen as critical RG in the cluster for both policies. The Critical RG can be a dummy RG without any resources. To see if a physical volume is SCSI Persistent Reserve Type 7H capable clmgr view pv <hdiskx> NAME="hdisk10" PVID="00f74e512845cbb7" UUID="16e2d679-e986-38e3-49d2-2650f3089bad" VOLUME_GROUP="None" TYPE="mpioosdisk" DESCRIPTION="MPIO 2810 XIV Disk" SIZE="16411" AVAILABLE="16411" CONCURRENT="true" Page 12 of 23
13 ENHANCED_CONCURRENT_MODE="true" STATUS="Available" SCSIPR_CAPABLE="Yes" Note: SCSIPR_CAPABLE=''No'' ( if a physical volume would not be scsi persistent reserve type 7H capable). Instead of checking for each disk one can run this command for each VG. To see if a volume group is SCSI Persistent Reserve Type 7H capable clmgr view vg <vg_name> NAME="vg1" TYPE="SCALABLE" NODES="powerha13,powerha14" LOGICAL_VOLUMES="" PHYSICAL_VOLUMES="hdisk77@powerha13@00f74e514901ede5" MIRROR_POOLS="" STRICT_MIRROR_POOLS="no" RESOURCE_GROUP="crg" AUTO_ACTIVATE="false" QUORUM="true" CONCURRENT_ACCESS="true" CRITICAL="false" ON_LOSS_OF_ACCESS="" NOTIFYMETHOD="'' MIGRATE_FAILED_DISKS="false" SYNCHRONIZE="false" LOGICAL_TRACK_GROUP_SIZE="512" MAX_PHYSICAL_PARTITIONS="32768" PPART_SIZE="16" MAX_LOGICAL_VOLUMES="256" MAJOR_NUMBER="56" IDENTIFIER="00f74e c eb564aee7" Page 13 of 23
14 TIMESTAMP="55af76ae130082e9" SCSIPR_CAPABLE="Yes" Note: SCSIPR_CAPABLE=''No'' ( if a volume group would not be scsi persistent reserve type 7H capable). Sometimes reserves are placed on the disks which may not allow to be changed to pr_shared. PowerHA SystemMirror will try to reset the policy to pr_shared, but it is recommended that customers make sure that there are no other applications changing the reserve policy of the disks. Reserve policy can be checked with lsattr command. # lsattr -El hdisk10 PCM PCM/friend/vscsi Path Control Module False algorithm fail_over Algorithm True... pvid 00c6fa22d39a8ec Physical volume identifier False queue_depth 3 Queue DEPTH True reserve_policy no_reserve Reserve Policy True+ PowerHA SystemMirror also provided a new command in 720 release to check the reserves on the disk. #clrsrvmgr -r -l hdisk1 -v Effective reserve policy on hdisk1 : no_reserve Configured reserve policy on hdisk1 Reservation status on /dev/hdisk1 : no_reserve : No Reservation This command can also be run on the VG directly #clrsrvmgr -r -g pgvg -v Effective reserve policy on hdisk1 : no_reserve Configured reserve policy on hdisk1 : no_reserve Reservation status on /dev/hdisk1 : No Reservation Page 14 of 23
15 Effective reserve policy on hdisk2 : no_reserve Configured reserve policy on hdisk2 : single_path Reservation status on /dev/hdisk2 : Single Path Reservation Effective reserve policy on hdisk3 : PR_shared Configured reserve policy on hdisk3 : PR_shared Reservation status on /dev/hdisk3 : SCSI PR Reservation This command can be run at any time on the cluster. Before the cluster services are started it can be run to make sure the state is no_reserve. (hdisk1 in the above example) When the cluster services are running we can make sure that the SCSI PR reservatio is set on the disk. (hdisk3 in the above example) Shared disks in the cluster should not be in single_path mode. (hdisk2 in the above example) Page 15 of 23
16 3.2.2 SMIT Panel PowerHA SystemMirror does not enable Disk Fencing quarantine policy by default. Administrator has to enable it before starting the cluster services. Note that this policy can not be enabled when cluster services are active on one or more nodes of the cluster. The policy cannot be changed on a VG that is already online. Note that this policy can be combined with the Tie Breaker split handling policies (Tie Breaker decisions are made first and then the Quarantine policy kicks in). Below are the SMIT screens used for configuring Disk Fencing. The menus can be accessed from Smit hacmp Custom Cluster Configuration Cluster Nodes and Networks Initial Cluster Setup (Custom) Configure Cluster Split and Merge Policy Quarantine Policy Disk Fencing or use the fastpath, smit cm_cluster_quarintine_disk_dialog. Page 16 of 23
17 Page 17 of 23
18 Page 18 of 23
19 3.2.3 clmgr Command Below is the clmgr option provided to configure Quarantine Policy clmgr modify cluster \ [ SPLIT_POLICY={none tiebreaker manual} ] \ [ TIEBREAKER=<disk> ] \ [ MERGE_POLICY={majority tiebreaker priority manual} ] \ [ NOTIFY_METHOD=<method> ] \ [ NOTIFY_INTERVAL=### ] \ [ MAXIMUM_NOTIFICATIONS=### ] \ [ DEFAULT_SURVIVING_SITE=<site> ] \ [ APPLY_TO_PPRC_TAKEOVER={yes no} ] [ ACTION_PLAN=reboot ] [ QUARANTINE_POLICY=<node_halt fencing halt_with_fencing>] \ [CRITICAL_RG=<rg_value>] Below are the clmgr options for managing Disk Fencing related operations To check if a hdisk supports SCSI 3 PR Type 7H clmgr query pv <hdiksxx> To check if all the disks of a volume group supports SCSI 3 PR Type 7H clmgr query vg <vg_name> To clear the SCSI3 reservation from a disk clmgr modify pv <hdiskxx> SCSIPR_ACTION=clear To clear the SCSI3 reservation from all the disk of a volume group clmgr modify vg <vg_name> SCSIPR_ACTION=clear Page 19 of 23
20 3.3 Runtime Considerations On a cluster where SCSI PR Disk fencing has been enabled we can check that the reservation state of the disk/vg using the clrsrvmgr command. A disk should have reserve policy as follows if cluster services are running and disk fencing is enabled. Configured Reserve Policy Effective Reserve Policy : PR_shared : PR_shared Reservation Status : SCSI PR reservation Write_Exclusive_All_Registrants Here, configured reserve policy is ODM reserve policy, effective reserve policy is kernel reserve policy and reservation status is device reservation state( Persistent Reserve Type). The output should be same on all nodes in the cluster where cluster services are active. In a cluster if a Split has happened and a node has been preempted from the disks, any IO that is happening on the disk will not be permitted any more. There will be LVM_IO_FAIl errors on the node and the Resource groups will eventually go into ERROR state. This happens on all the nodes on the loosing island. The Resource Groups cannot be recovered on these node without stopping the cluster services (Even when the split has been healed) This restriction is in place so a resource cannot be started on this node again and there by causing any data corruption. Once the cluster has healed the nodes on the loosing side should stop cluster services and start them again to rejoin the cluster. On a Linked or Stretched cluster with some split and merge Policies setup the nodes on the loosing island will reboot, so cluster services need to be started after the Split has been healed. Page 20 of 23
21 If there ever was a scenario which happens due to timing problems that the active node end up in a scenario where all the Vgs cannot be brought online because of Disk fencing issues the following steps can be followed to recover the reservations on that node. Smit hacmp Problem Determination Tools And we have to selct the RG in Error in the next screen. This will clear any RG's in Error state and the RG can be brought online again. Page 21 of 23
22 4 References scdisk SCSI Device Driver T10 Documents a. SPC-4 b. IBMer PPT: Understanding Persistent Reserves 01.ibm.com/support/knowledgecenter/SSFKCN_4.1.0/com.ibm.cluster.gpfs.v4r1.gpf s500.doc/bl1pdg_understandpr.htm SCSI reservation methodologies Page 22 of 23
23 Disclaimers: This article is to provide an overview of the Disk Fencing management and is not expected to be complete. Refer to PowerHA SystemMirror documentation for the most recent and correct information about SCSI-3 Disk Fencing. All the information in this article is opinions from the authors and no way represents the position of PowerHA SystemMirror product or IBM. Page 23 of 23
IBM PowerHA SystemMirror for AIX. Standard Edition. Version 7.1. PowerHA SystemMirror concepts IBM
IBM PowerHA SystemMirror for AIX Standard Edition Version 7.1 PowerHA SystemMirror concepts IBM IBM PowerHA SystemMirror for AIX Standard Edition Version 7.1 PowerHA SystemMirror concepts IBM Note Before
More informationThe Host Server. AIX Configuration Guide. August The Data Infrastructure Software Company
The Host Server AIX Configuration Guide August 2017 This guide provides configuration settings and considerations for SANsymphony Hosts running IBM's AIX. Basic AIX administration skills are assumed including
More informationNotes for migrating to EMC or Shark with Fibre Channel boot. Planning Notes
Notes for migrating to EMC or Shark with Fibre Channel boot R. Ballough, 1/15/2004 Planning Notes 1. Reference document SANBootQA0103.pdf for a complete list of requirements when booting from SAN attached
More informationPower Systems High Availability & Disaster Recovery
Power Systems High Availability & Disaster Recovery Solutions Comparison of various HA & DR solutions for Power Systems Authors: Carl Burnett, Joe Cropper, Ravi Shankar Table of Contents 1 Abstract...
More informationIBM System Storage N series
IBM System Storage N series Host Settings Affected by AIX Host Utilities The AIX Host Utilities sometimes recommend that settings on the host be modified. These recommendations are based on research and
More informationIBM PowerHA SystemMirror for Linux. Version IBM
IBM PowerHA SystemMirror for Linux Version 7.2.2 IBM IBM PowerHA SystemMirror for Linux Version 7.2.2 IBM Note Before using this information and the product it supports, read the information in Notices
More informationGrover Davidson Senior Software Engineer, AIX Development Support IBM pin541 Hidden Features and Functionality in AIX
Grover Davidson grover@us.ibm.com Senior Software Engineer, AIX Development Support IBM pin541 Hidden Features and Functionality in AIX Copyright IBM Corporation 2014 Technical University/Symposia materials
More informationPowerHA SystemMirror 6.1 to migrating Prerequisites
PowerHA SystemMirror 6.1 to 7.1.3 migrating Prerequisites Abstract: Whenever we are going to perform PowerHA SystemMirror rolling or snapshot migration from 6.1 to 7.1.3 we have to fulfill the prerequisites.
More informationInstallation Manual. NEXSAN MSIO for AIX. Version 2.1
NEXSAN MSIO for AIX Installation Manual Version 2.1 NEXSAN 555 St. Charles Drive, Suite 202, Thousand Oaks, CA 91360 p. 866.4.NEXSAN f. 866.418.2799 COPYRIGHT Copyright 2009 2011 by Nexsan Corporation.
More informationPowerHA Enterprise Edition Cross Reference
PowerHA Enterprise Edition Cross Reference Replication Option HA version (minimum) AIX 5.3 AIX 6.1 AIX 7.1 VIOS NPIV GLVM synchronous 5.2+IY66555 ML2 GLVM asynchronous 5.5 SP1 NO Yes-TL2 SP3 SVC Metro
More informationFlexArray Virtualization Implementation Guide for Third- Party Storage
ONTAP 9 FlexArray Virtualization Implementation Guide for Third- Party Storage June 2018 215-11150_F0 doccomments@netapp.com Table of Contents 3 Contents Where to find information for configurations with
More informationIBM Power Systems with POWER7 and AIX Technical Sales Skills - v1.
IBM 000-107 Power Systems with POWER7 and AIX Technical Sales Skills - v1 http://killexams.com/exam-detail/000-107 A customer has applications that run on POWER4 servers that require AIX 5.2. The customer
More informationCLUSTERING. What is Clustering?
What is Clustering? CLUSTERING A cluster is a group of independent computer systems, referred to as nodes, working together as a unified computing resource. A cluster provides a single name for clients
More informationThe Contents and Structure of this Manual. This document is composed of the following 12 chapters.
Preface This document briefly explains the operations that need to be performed by the user in order to connect an ETERNUS2000 model 100 or 200, ETERNUS4000 model 300, 400, 500, or 600, or ETERNUS8000
More informationHitachi Dynamic Link Manager (for AIX) Release Notes
Hitachi Dynamic Link Manager (for AIX) 8.4.1-00 Release Notes Contents About this document... 1 Intended audience... 1 Getting help... 2 About this release... 2 Product package contents... 2 New features
More informationEMC VPLEX Geo with Quantum StorNext
White Paper Application Enabled Collaboration Abstract The EMC VPLEX Geo storage federation solution, together with Quantum StorNext file system, enables a global clustered File System solution where remote
More informationVERITAS Dynamic MultiPathing (DMP) Increasing the Availability and Performance of the Data Path
White Paper VERITAS Storage Foundation for Windows VERITAS Dynamic MultiPathing (DMP) Increasing the Availability and Performance of the Data Path 12/6/2004 1 Introduction...3 Dynamic MultiPathing (DMP)...3
More informationFlexArray Virtualization
Updated for 8.3.1 FlexArray Virtualization Implementation Guide for Third-Party Storage NetApp, Inc. 495 East Java Drive Sunnyvale, CA 94089 U.S. Telephone: +1 (408) 822-6000 Fax: +1 (408) 822-4501 Support
More informationUsing the Geographic LVM in AIX 5L
Using the Geographic LVM in AIX 5L September 2005 Steve Tovcimak IBM Corporation Table of Contents Overview...3 Basic Concepts...3 GLVM Utilities...7 Quorum and Remote Physical Volume Failure...8 Avoiding
More informationDisk I/O and the Network
Page 1 of 5 close window Print Disk I/O and the Network Increase performance with more tips for AIX 5.3, 6.1 and 7 October 2010 by Jaqui Lynch Editor s Note: This is the concluding article in a two-part
More informationRed Hat Clustering: Best Practices & Pitfalls. Lon Hohberger Principal Software Engineer Red Hat May 2013
Red Hat Clustering: Best Practices & Pitfalls Lon Hohberger Principal Software Engineer Red Hat May 2013 1 Red Hat Clustering: Best Practices & Pitfalls Why Cluster? I/O Fencing and Your Cluster 2-Node
More informationWelcome to PowerPath Foundations
IMPACT Home IMPACT modules consist of focused, in-depth training content that can be consumed in about 1-2 hours Welcome to PowerPath Foundations Course Description Start Training Run/Download the PowerPoint
More informationPower Systems SAN Multipath Configuration Using NPIV v1.2
v1.2 Bejoy C Alias IBM India Software Lab Revision History Date of this revision: 27-Jan-2011 Date of next revision : TBD Revision Number Revision Date Summary of Changes Changes marked V1.0 23-Sep-2010
More information1 BRIEF / Oracle Solaris Cluster Features and Benefits
Oracle Solaris Cluster is a comprehensive high availability (HA) and disaster recovery (DR) solution for Oracle SPARC and x86 environments that is based on Oracle Solaris. It combines extreme service availability
More informationVeritas Cluster Server 6.0
Veritas Cluster Server 6.0 New Features and Capabilities Anthony Herr Product Manager - Availability Products What does function does VCS perform? High Availability Ensure an application, in either a physical
More informationStarWind Virtual SAN Windows Geo-Clustering: SQL Server
#1 HyperConverged Appliance for SMB and ROBO StarWind Virtual SAN Windows Geo-Clustering: SQL Server FEBRUARY 2016 TECHNICAL PAPER EDWIN SARMIENTO, Microsoft SQL Server MVP, Microsoft Certified Master
More informationTivoli System Automation for Multiplatforms Version 4 Release 1. Administrator's and User's Guide IBM SC
Tivoli System Automation for Multiplatforms Version 4 Release 1 Administrator's and User's Guide IBM SC34-2698-04 Tivoli System Automation for Multiplatforms Version 4 Release 1 Administrator's and User's
More informationExpert Reference Series of White Papers. Understanding the AIX Object Data Manager
Expert Reference Series of White Papers Understanding the AIX Object Data Manager 1-800-COURSES www.globalknowledge.com Understanding the AIX Object Data Manager Iain Campbell, UNIX/Linux Open Systems
More informationCompellent Storage Center
How To Setup a Microsoft Windows Server 2003 Failover Cluster Compellent Corporate Office Compellent Technologies 7625 Smetana Lane Eden Prairie, Minnesota 55344 www.compellent.com Contents Contents...
More informationWHITE PAPER: ENTERPRISE SOLUTIONS. Veritas Storage Foundation for Windows Dynamic Multi-pathing Option. Competitive Comparisons
WHITE PAPER: ENTERPRISE SOLUTIONS Veritas Storage Foundation for Windows Competitive Comparisons White Paper: Enterprise Solutions Veritas Storage Foundation for Windows Contents Introduction........................................................................4
More informationEMC PowerPath for AIX Version 5.1
EMC PowerPath for AIX Version 5.1 Installation and Administration Guide P/N 300-005-461 REV A03 EMC Corporation Corporate Headquarters: Hopkinton, MA 01748-9103 1-508-435-1000 www.emc.com Copyright 1997-2008
More informationThe Clone Task Force Local Replication, CLARiiON Environment
The Clone Task Force Local Replication, CLARiiON Environment EMC Proven Professional Knowledge Sharing May, 2007 Fernando Moreno Liso System Executive Engineer CLARiiON Specialist Comparex Espana, S.A
More informationWhat's in this guide... 4 Documents related to NetBackup in highly available environments... 5
Contents Chapter 1 About in this guide... 4 What's in this guide... 4 Documents related to NetBackup in highly available environments... 5 Chapter 2 NetBackup protection against single points of failure...
More informationJunos Security. Chapter 11: High Availability Clustering Implementation
Junos Security Chapter 11: High Availability Clustering Implementation 2012 Juniper Networks, Inc. All rights reserved. www.juniper.net Worldwide Education Services Chapter Objectives After successfully
More informationEMC VPLEX with Quantum Stornext
White Paper Application Enabled Collaboration Abstract The EMC VPLEX storage federation solution together with Quantum StorNext file system enables a stretched cluster solution where hosts has simultaneous
More informationFUJITSU Storage ETERNUS Multipath Driver 2 (for AIX) Installation Information
FUJITSU Storage ETERNUS Multipath Driver 2 (for AIX) Installation Information Oct 2016 Contents Contents... 1 Supported Operating System (OS) Versions... 1 Supported Disk Storage Systems... 1 Connection
More informationVMware vsphere Metro Storage Cluster Recommended Practices November 08, 2017
VMware vsphere Metro Storage Cluster Recommended Practices November 08, 2017 1 Table of Contents 1. Purpose and Overview 1.1.Purpose and Overview 1.2.Target Audience 2. Introduction 2.1.Introduction 2.2.Technical
More informationRelease Notes P/N Rev A02
EMC PowerPath for Windows Version 4.6 and Point Releases Release Notes P/N 300-003-931 Rev A02 November 27, 2006 These release notes contain information about features, system requirements, known problems,
More informationBalancing RTO, RPO, and budget. Table of Contents. White Paper Seven steps to disaster recovery nirvana for wholesale distributors
White Paper Seven steps to disaster recovery nirvana for wholesale distributors Balancing RTO, RPO, and budget In our last white paper, Thinking through the unthinkable: Disaster recovery for wholesale
More informationIBM p5 and pseries Enterprise Technical Support AIX 5L V5.3. Download Full Version :
IBM 000-180 p5 and pseries Enterprise Technical Support AIX 5L V5.3 Download Full Version : https://killexams.com/pass4sure/exam-detail/000-180 A. The LPAR Configuration backup is corrupt B. The LPAR Configuration
More informationNumber: Passing Score: 800 Time Limit: 120 min File Version: 1.0. Vendor: IBM. Exam Code:
000-332 Number: 000-000 Passing Score: 800 Time Limit: 120 min File Version: 1.0 http://www.gratisexam.com/ Vendor: IBM Exam Code: 000-332 Exam Name: High Availability for AIX - Technical Support and Administration
More informationSession Title: Designing a PowerHA SystemMirror for AIX Disaster Recovery Solution
IBM Power Systems Technical University October 18 22, 2010 Las Vegas, NV Session Title: Designing a PowerHA SystemMirror for AIX Disaster Recovery Solution Session ID: HA18 (AIX) Speaker Name: Michael
More informationDynamic Multi-Pathing 7.0 Administrator's Guide - AIX
Dynamic Multi-Pathing 7.0 Administrator's Guide - AIX December 2015 Dynamic Multi-Pathing Administrator's Guide The software described in this book is furnished under a license agreement and may be used
More informationExam Name: High Availability for AIX - Technical Support
Exam Code: 000-102 Exam Name: High Availability for AIX - Technical Support and Administration Vendor: IBM Version: DEMO Part: A 1: A customer is in the process of testing their configuration prior to
More informationLive Partition Mobility Update
Power Systems ATS Live Partition Mobility Update Ron Barker Power Advanced Technical Sales Support Dallas, TX Agenda Why you should be planning for partition mobility What are your options? Which is best
More informationSM B10: Rethink Disaster Recovery: Replication and Backup Are Not Enough
SM B10: Rethink Disaster Recovery: Replication and Backup Are Not Enough Paul Belk Director, Product Management Mike Weiss Staples Ranga Rajagopalan Principal Product Manager Tsunami Hurricane Philippines
More informationMicrosoft Exam Questions & Answers
Microsoft 70-483 Exam Questions & Answers Number: 70-483 Passing Score: 800 Time Limit: 120 min File Version: 12.8 http://www.gratisexam.com/ Microsoft 70-483 Exam Questions & Answers Exam Name: Programming
More informationHigh Availability for Oracle Database with IBM PowerHA SystemMirror and IBM Spectrum Virtualize HyperSwap
Front cover High Availability for Oracle Database with IBM PowerHA SystemMirror and IBM Spectrum Virtualize HyperSwap Ian MacQuarrie Redpaper High Availability for Oracle Database with IBM PowerHA SystemMirror
More informationGeographic LVM: Planning and administration guide
High Availability Cluster Multi-Processing XD (Extended Distance) Geographic LVM: Planning and administration guide SA23-1338-07 High Availability Cluster Multi-Processing XD (Extended Distance) Geographic
More informationThis five-day, instructor-led, hands-on class covers how to use Veritas Cluster Server to manage applications in a high availability environment.
Veritas Cluster Server 6.0 for UNIX: Administration Day(s): 5 Course Code: HA0434 Overview The Veritas Cluster Server 6.0 for UNIX: Administration course is designed for the IT professional tasked with
More informationAIX Host Utilities 6.0 Installation and Setup Guide
IBM System Storage N series AIX Host Utilities 6.0 Installation and Setup Guide GC27-3925-01 Table of Contents 3 Contents Preface... 6 Supported features... 6 Websites... 6 Getting information, help,
More informationHow to Implement High Availability for the SAS Metadata Server Using High Availability Cluster Multi-Processing (HACMP)
Technical Paper How to Implement High Availability for the SAS Metadata Server Using High Availability Cluster Multi-Processing (HACMP) Technical White Paper by SAS and IBM Table of Contents Abstract...
More informationBroker Clusters. Cluster Models
4 CHAPTER 4 Broker Clusters Cluster Models Message Queue supports the use of broker clusters: groups of brokers working together to provide message delivery services to clients. Clusters enable a Message
More informationIBM POWERVM WITH FLASHARRAY
IBM POWERVM WITH FLASHARRAY White paper - September 2017 Contents Executive Summary... 3 Introduction... 3 Pure Storage FlashArray... 3 IBM Power System... 3 DEVICES Virtualization With IBM PowerVM...
More informationThis page is intentionally left blank.
This page is intentionally left blank. Preface This ETERNUS Multipath Driver User's Guide describes the features, functions, and operation of the "ETERNUS Multipath Driver" (hereafter referred to as "Multipath
More informationIntroduction to PowerHA SystemMirror for AIX V 7.1 Managed with IBM Systems Director
Introduction to PowerHA SystemMirror for AIX V 7.1 Managed with Director IBM s High Availability Software for POWER Based Systems Glenn Miller Certified IT Specialist Systems Software Architect gemiller@us.ibm.com
More informationOverview. CPS Architecture Overview. Operations, Administration and Management (OAM) CPS Architecture Overview, page 1 Geographic Redundancy, page 5
CPS Architecture, page 1 Geographic Redundancy, page 5 CPS Architecture The Cisco Policy Suite (CPS) solution utilizes a three-tier virtual architecture for scalability, system resilience, and robustness
More informationIBM Version 7 Release 3. Easy Tier Server SC
IBM Version 7 Release 3 Easy Tier Server SC27-5430-02 This edition applies to Version 7, Release 3 of the IBM Easy Tier Server and to all subsequent releases and modifications until otherwise indicated
More informationSVC VOLUME MIGRATION
The information, tools and documentation ( Materials ) are being provided to IBM customers to assist them with customer installations. Such Materials are provided by IBM on an as-is basis. IBM makes no
More informationDB2 purescale Active/Active High Availability is Here!
purescale Active/Active High Availability is Here! Session C04, for LUW Aamer Sachedina STSM, IBM Toronto Lab November 9, 2010, 8:30am 0 purescale is state of the art for LUW technology which offers active/active
More informationConfiguration Guide for IBM AIX Host Attachment Hitachi Virtual Storage Platform Hitachi Universal Storage Platform V/VM
Configuration Guide for IBM AIX Host Attachment Hitachi Virtual Storage Platform Hitachi Universal Storage Platform V/VM FASTFIND LINKS Document Organization Product Version Getting Help Contents MK-96RD636-05
More informationInstalling the IBM ServeRAID Cluster Solution
Installing the IBM ServeRAID Cluster Solution For IBM Netfinity ServeRAID-4x Ultra160 SCSI Controllers Copyright IBM Corp. 2000 1 2 IBM Netfinity Installing the IBM ServeRAID Cluster Solution Chapter 1.
More informationFlexArray Virtualization
Updated for 8.3.2 FlexArray Virtualization Installation Requirements and Reference Guide NetApp, Inc. 495 East Java Drive Sunnyvale, CA 94089 U.S. Telephone: +1 (408) 822-6000 Fax: +1 (408) 822-4501 Support
More informationAbstract /10/$26.00 c 2010 IEEE
Abstract Clustering solutions are frequently used in large enterprise and mission critical applications with high performance and availability requirements. This is achieved by deploying multiple servers
More informationECS High Availability Design
ECS High Availability Design March 2018 A Dell EMC white paper Revisions Date Mar 2018 Aug 2017 July 2017 Description Version 1.2 - Updated to include ECS version 3.2 content Version 1.1 - Updated to include
More informationUnderstanding high availability with WebSphere MQ
Mark Hiscock Software Engineer IBM Hursley Park Lab United Kingdom Simon Gormley Software Engineer IBM Hursley Park Lab United Kingdom May 11, 2005 Copyright International Business Machines Corporation
More information3.1. Storage. Direct Attached Storage (DAS)
3.1. Storage Data storage and access is a primary function of a network and selection of the right storage strategy is critical. The following table describes the options for server and network storage.
More informationNew England Data Camp v2.0 It is all about the data! Caregroup Healthcare System. Ayad Shammout Lead Technical DBA
New England Data Camp v2.0 It is all about the data! Caregroup Healthcare System Ayad Shammout Lead Technical DBA ashammou@caregroup.harvard.edu About Caregroup SQL Server Database Mirroring Selected SQL
More informationPreventing Silent Data Corruption Using Emulex Host Bus Adapters, EMC VMAX and Oracle Linux. An EMC, Emulex and Oracle White Paper September 2012
Preventing Silent Data Corruption Using Emulex Host Bus Adapters, EMC VMAX and Oracle Linux An EMC, Emulex and Oracle White Paper September 2012 Preventing Silent Data Corruption Introduction... 1 Potential
More informationNamenode HA. Sanjay Radia - Hortonworks
Namenode HA Sanjay Radia - Hortonworks Sanjay Radia - Background Working on Hadoop for the last 4 years Part of the original team at Yahoo Primarily worked on HDFS, MR Capacity scheduler wire protocols,
More informationEMC Simple Support Matrix
EMC Simple Support Matrix EMC Symmetrix DMX-3 and DMX-4 OCTOBER 2014 P/N 300-013-399 REV 33 2011-2014 EMC Corporation. All Rights Reserved. EMC believes the information in this publication is accurate
More informationVERITAS Dynamic Multipathing. Increasing the Availability and Performance of the Data Path
VERITAS Dynamic Multipathing Increasing the Availability and Performance of the Data Path 1 TABLE OF CONTENTS I/O Path Availability and Performance... 3 Dynamic Multipathing... 3 VERITAS Storage Foundation
More informationStorage agnostic end to end storage information for long distance high availability. Vijay Kumar Shankarappa Rupesh Thota IBM India
Storage agnostic end to end storage information for long distance high availability Vijay Kumar Shankarappa Rupesh Thota IBM India Contents 1) High availability/recovery solutions 2) Long distance availability
More informationDynamic Multi-Pathing 7.2 Administrator's Guide - AIX
Dynamic Multi-Pathing 7.2 Administrator's Guide - AIX October 2016 Dynamic Multi-Pathing Administrator's Guide Last updated: 2016-10-31 Document version: 7.2 Rev 0 Legal Notice Copyright 2016 Veritas Technologies
More informationSimplified Storage Migration for Microsoft Cluster Server
Simplified Storage Migration for Microsoft Cluster Server Using VERITAS Volume Manager for Windows 2000 with Microsoft Cluster Server V E R I T A S W H I T E P A P E R June 2001 Table of Contents Overview...................................................................................1
More informationIBM EXAM QUESTIONS & ANSWERS
IBM 000-106 EXAM QUESTIONS & ANSWERS Number: 000-106 Passing Score: 800 Time Limit: 120 min File Version: 38.8 http://www.gratisexam.com/ IBM 000-106 EXAM QUESTIONS & ANSWERS Exam Name: Power Systems with
More informationStep-by-Step Guide to Installing Cluster Service
Page 1 of 23 TechNet Home > Products & Technologies > Windows 2000 Server > Deploy > Configure Specific Features Step-by-Step Guide to Installing Cluster Service Topics on this Page Introduction Checklists
More informationSynology High Availability (SHA)
Synology High Availability (SHA) Based on DSM 5.1 Synology Inc. Synology_SHAWP_ 20141106 Table of Contents Chapter 1: Introduction... 3 Chapter 2: High-Availability Clustering... 4 2.1 Synology High-Availability
More informationIBM Exam A Virtualization Technical Support for AIX and Linux Version: 6.0 [ Total Questions: 93 ]
s@lm@n IBM Exam A4040-101 Virtualization Technical Support for AIX and Linux Version: 6.0 [ Total Questions: 93 ] IBM A4040-101 : Practice Test Question No : 1 Which of the following IOS commands displays
More informationDeployment Guide for SRX Series Services Gateways in Chassis Cluster Configuration
Deployment Guide for SRX Series Services Gateways in Chassis Cluster Configuration Version 1.2 June 2013 Juniper Networks, 2013 Contents Introduction... 3 Chassis Cluster Concepts... 4 Scenarios for Chassis
More informationCritical Resource Analysis (CRA) White Paper
Critical Resource Analysis (CRA) White Paper SUMMARY... 3 Introduction to Critical Resource Analysis... 4 CRA return values... 4 Mass Storage CRA Scenarios... 5 Boot Path Configuration Scenarios... 5 Scenario
More informationvsan Remote Office Deployment January 09, 2018
January 09, 2018 1 1. vsan Remote Office Deployment 1.1.Solution Overview Table of Contents 2 1. vsan Remote Office Deployment 3 1.1 Solution Overview Native vsphere Storage for Remote and Branch Offices
More informationEssentials. Oracle Solaris Cluster. Tim Read. Upper Saddle River, NJ Boston Indianapolis San Francisco. Capetown Sydney Tokyo Singapore Mexico City
Oracle Solaris Cluster Essentials Tim Read PRENTICE HALL Upper Saddle River, NJ Boston Indianapolis San Francisco New York Toronto Montreal London Munich Paris Madrid Capetown Sydney Tokyo Singapore Mexico
More informationAIX5 Initial Settings for Databases Servers
Introduction Here are the AIX 5L settings I automatically change when installing a pseries database server. They are provided here as a reference point for tuning an AIX system. As always, all settings
More informationClustering and Storage Management In Virtualized Environments Rasmus Rask Eilersen
Clustering and Storage Management In Virtualized Environments Rasmus Rask Eilersen Principal Systems Engineer 1 Tak til vores sponsorer Technology Days 2013 2 VIRTUALIZATION GROWTH 1 new VM every 6 seconds
More informationDell EMC ME4 Series Storage Systems. Release Notes
Dell EMC ME4 Series Storage Systems Release Notes Notes, cautions, and warnings NOTE: A NOTE indicates important information that helps you make better use of your product. CAUTION: A CAUTION indicates
More informationHitachi Dynamic Link Manager (for AIX) v Release Notes
Hitachi Dynamic Link Manager (for AIX) v8.2.1-00 Release Notes Contents About this document... 1 Intended audience... 1 Getting help... 2 About this release... 2 Product package contents... 2 New features
More informationHP StoreVirtual Storage Multi-Site Configuration Guide
HP StoreVirtual Storage Multi-Site Configuration Guide Abstract This guide contains detailed instructions for designing and implementing the Multi-Site SAN features of the LeftHand OS. The Multi-Site SAN
More informationPowerPath PRODUCT GUIDE. Version 4.6 P/N REV A03
PowerPath Version 4.6 PRODUCT GUIDE P/N 300-003-927 REV A03 EMC Corporation Corporate Headquarters: Hopkinton, MA 01748-9103 1-508 -435-1000 www.emc.com Copyright 1997-2006 EMC Corporation. All rights
More informationManaging VMAX. Summary of Steps. This chapter contains the following sections:
This chapter contains the following sections: Summary of Steps, page 1 VMAX Management, page 2 Thin Pools, page 2 Data Devices, page 3 Thin Devices, page 4 Meta Devices, page 6 Initiator Groups, page 7
More informationInfrastructure Provisioning with System Center Virtual Machine Manager
Infrastructure Provisioning with System Center Virtual Machine Manager Course Details Duration: Course code: 5 Days M10981 Overview: Learn how to install and configure Microsoft System Center 2012 R2 Virtual
More informationNative vsphere Storage for Remote and Branch Offices
SOLUTION OVERVIEW VMware vsan Remote Office Deployment Native vsphere Storage for Remote and Branch Offices VMware vsan is the industry-leading software powering Hyper-Converged Infrastructure (HCI) solutions.
More informationIBM Solutions Advanced Technical Support
DB2 and SAP Disaster Recovery using DS8300 Global Mirror IBM Solutions Advanced Technical Support Nasima Ahmad Chris Eisenmann Jean-Luc Degrenand Mark Gordon Mark Keimig Damir Rubic Version: 1.1 Date:
More informationEMC VPLEX WITH SUSE HIGH AVAILABILITY EXTENSION BEST PRACTICES PLANNING
EMC VPLEX WITH SUSE HIGH AVAILABILITY EXTENSION BEST PRACTICES PLANNING ABSTRACT This White Paper provides a best practice to install and configure SUSE SLES High Availability Extension (HAE) with EMC
More informationHigh Availability Procedures and Guidelines
IBM FileNet Image Services Version 4.2 High Availability Procedures and Guidelines SC19-3303-0 Contents About this manual 11 Audience 11 Document revision history 11 Accessing IBM FileNet Documentation
More informationVERITAS Volume Manager for Windows 2000 VERITAS Cluster Server for Windows 2000
WHITE PAPER VERITAS Volume Manager for Windows 2000 VERITAS Cluster Server for Windows 2000 VERITAS CAMPUS CLUSTER SOLUTION FOR WINDOWS 2000 WHITEPAPER 1 TABLE OF CONTENTS TABLE OF CONTENTS...2 Overview...3
More informationHACMP Smart Assist for Oracle User s Guide
High Availability Cluster Multi-Processing for AIX 5L HACMP Smart Assist for Oracle User s Guide Version 5.3 SC23-5178-01 Second Edition (August 2005) Before using the information in this book, read the
More informationInfoSphere Warehouse with Power Systems and EMC CLARiiON Storage: Reference Architecture Summary
InfoSphere Warehouse with Power Systems and EMC CLARiiON Storage: Reference Architecture Summary v1.0 January 8, 2010 Introduction This guide describes the highlights of a data warehouse reference architecture
More informationIBM Storwize V7000 with IBM PowerHA SystemMirror
IBM Storwize V7000 with IBM PowerHA SystemMirror Proof of concept and configuration Zane Russell IBM Systems and Technology Group ISV Enablement March 2011 Copyright IBM Corporation, 2011. Abstract...
More informationVeritas Cluster Server 6.0
Veritas Cluster Server 6.0 New Features and Capabilities SF 6.0 Features Deep Dive Anthony Herr, Eric Hennessey SAMG Technical Product Management What does VCS do for me? High Availability Ensure an application,
More information