Power Systems SAN Multipath Configuration Using NPIV v1.2

Similar documents
Experiences with VIOS support for IBM i

PowerVM - Dynamically adding a Virtual Fibre Channel adapter to a client partition

Notes for migrating to EMC or Shark with Fibre Channel boot. Planning Notes

IBM POWERVM WITH FLASHARRAY

Application and Partition Mobility

The Contents and Structure of this Manual. This document is composed of the following 12 chapters.

IBM i on a Power Blade Read-me First. Vess Natchev Power Blades Virtualization Cloud IBM Rochester, MN

BMC Capacity Optimization Extended Edition AIX Enhancements

Technical Note P/N REV A01 November 24, 2008

Advanced Virtual I/O Server Configurations. César Diniz Maciel Consulting IT Specialist IBM US

Live Partition Mobility Update

Exam Name: XIV Replication and Migration V1

Disk I/O and the Network

IBM Virtualization Technical Support for AIX and Linux - v2.

IBM i Virtual I/O Performance in an IBM System Storage SAN Volume Controller with IBM System Storage DS8000 Environment

Host Attachment Guide

Live Partition Mobility

How to Configure NPIV in a Microsoft Hyper-V Environment

IBM Exam A Virtualization Technical Support for AIX and Linux Version: 6.0 [ Total Questions: 93 ]

SVC VOLUME MIGRATION

IBM EXAM QUESTIONS & ANSWERS

C exam. Number: C Passing Score: 800 Time Limit: 120 min IBM C IBM AIX Administration V1.

Dell EMC SC Series Storage with SAS Front-end Support for VMware vsphere

IBM XIV Host Attachment Kit for AIX Version Release Notes

Integrated Virtualization Manager. ESCALA Power7 REFERENCE 86 A1 41FF 08

IBM IBM Storage Networking Solutions Version 1.

PrepAwayExam. High-efficient Exam Materials are the best high pass-rate Exam Dumps

Common 2009 IBM I. Miroslav Iwachow IT Architect IBM CEE

Technical Brief: How to Configure NPIV on VMware vsphere 4.0

Power Systems with POWER8 Scale-out Technical Sales Skills V1

IBM XIV Host Attachment Kit for AIX. Version Release Notes. First Edition (September 2011)

AIX Host Utilities 6.0 Installation and Setup Guide

IBM SYSTEM POWER7. PowerVM. Jan Kristian Nielsen Erik Rex IBM Corporation

IBM p5 and pseries Technical Sales Support. Download Full Version :

Planning and Implementing NPIV For System z

IBM Storage Networking Solutions Version 2. Download Full Version :

A. Both Node A and Node B will remain on line and all disk shelves will remain online.

Virtualization Technical Support for AIX and Linux - v2

PowerHA Enterprise Edition Cross Reference

IBM Systems Director VMControl

The Power of PowerVM Power Systems Virtualization. Eyal Rubinstein

IBM XIV Host Attachment Kit for AIX Version Release Notes

Fibre Channel and iscsi Configuration Guide for the Data ONTAP 8.0 Release Family

HS22, HS22v, HX5 Boot from SAN with QLogic on IBM UEFI system.

IBM IBM Open Systems Storage Solutions Version 4. Download Full Version :

Veritas InfoScale 7.0 Virtualization Guide - AIX

IBM XIV Host Attachment Kit for AIX Version Release Notes IBM

Software Installation Reference

Vendor: HuaWei. Exam Code: H Exam Name: HCNP-Storage-CUSN(Constructing Unifying Storage Network) Version: Demo

EXAM - NS NetApp Certified Implementation Engineer - SAN, Clustered Data ONTAP. Buy Full Product.

41CO Operation and Maintaining PowerVM VIOS with IBM i

Exam Name: Midrange Storage Technical Support V2

IBM Version 7 Release 3. Easy Tier Server SC

Copyright International Business Machines Corporation 2008, 2009, 2010, 2011, 2012, 2013 All rights reserved.

Vendor: IBM. Exam Code: C Exam Name: Power Systems with POWER8 Scale-out Technical Sales Skills V1. Version: Demo

IBM. Multipath Subsystem Device Driver User s Guide. IBM System Storage GC

IBM System Storage SAN Volume Controller (SVC)

Host and storage system rules

Configuration Guide -Server Connection-

Configuration Guide -Server Connection-

SAN Configuration Guide

Vendor: IBM. Exam Code: C Exam Name: Enterprise Technical Support for AIX and Linux -v2. Version: Demo

Exam : S Title : Snia Storage Network Management/Administration. Version : Demo

Storage Protocol Offload for Virtualized Environments Session 301-F

Network Administrator s Guide to Actifio Copy Data Management

IT Certification Exams Provider! Weofferfreeupdateserviceforoneyear! h ps://

POWER Block Course Assignment 2: System Management. Hasso-Plattner Institute

COPYRIGHTED MATERIAL. Windows Server 2008 Storage Services. Chapter. in this chapter:

SAN Volume Controller (SVC) V6 Planning and Implementation Workshop

NS0-506 Q&As NetApp Certified Implementation Engineer - SAN, Clustered Data ONTAP

Virtualization Benefits IBM Corporation

SAN Implementation (SANIW)

... Oracle Database 11g Release 2 with Oracle Real Application Clusters on IBM Flex System tm p460 Compute Nodes with IBM Storwize tm V7000 Unified

Dell EMC SC Series Storage: Microsoft Multipath I/O

IBM ^ BladeCenter Remote SAN Boot for HS20 and HS40

Deployment Guide: Network Convergence with Emulex OneConnect FCoE CNA and Windows Server Platform

F port channel trunk from MDS 9148 (NPV) to MDS 9509 (NPIV) Configuration Example

Exam Name: Virtualization Technical Support for AIX and

FUJITSU Storage ETERNUS Multipath Driver 2 (for AIX) Installation Information

IBM. Availability Implementing high availability. IBM i 7.1

The Contents and Structure of this Manual. This document is composed of the following six chapters and three appendices.

Actual4Test. Actual4test - actual test exam dumps-pass for IT exams

The Contents and Structure of this Manual. This document is composed of the following four chapters and an appendix.

networkserversuite AIX & AIX VIOS Agent Installation

PASS4TEST. IT Certification Guaranteed, The Easy Way! We offer free update service for one year

MAGiC Conference September, 2017

Exam Name: Enterprise Technical Support for AIX and

A Mainframe Guy Discovers Blades..as in zenterprise Blade Extension

Virtual I/O Server Beyond the Basics

Learn Your Alphabet - SRIOV, NPIV, RoCE, iwarp to Pump Up Virtual Infrastructure Performance

IBM Emulex 16Gb Fibre Channel HBA Evaluation

S SNIA Storage Networking Management & Administration

Server Support Matrix ETERNUS Disk storage systems Server Connection Guide (Fibre Channel) for Oracle Solaris

Mr. D. Vinod IT Architect

A TPC Standby Solution User's Guide

IBM System Storage Support for Microsoft Volume Shadow Copy Service and Virtual Disk Service Version User Guide SC

IBM Spectrum Protect Snapshot for Custom Applications Version Installation and User Guide UNIX and Linux IBM

3.1. Storage. Direct Attached Storage (DAS)

IBM Tivoli Storage FlashCopy Manager for Custom Applications Version Installation and User Guide UNIX and Linux IBM

Large SAN Design Best Practices Using Cisco MDS 9700 and MDS 9500 Multilayer Directors

Transcription:

v1.2 Bejoy C Alias IBM India Software Lab

Revision History Date of this revision: 27-Jan-2011 Date of next revision : TBD Revision Number Revision Date Summary of Changes Changes marked V1.0 23-Sep-2010 1. Initial creation N V1.1 27-Oct-2010 1. Included bookmarks for figures 2. Modified section 2.4 to include details about Live Partition Mobility (Thanks to Bob Foster for pointing out this missing part) V1.2 27-Jan-2011 1. Updated section 2.3.2 with review comments from Dan Braden (IBM US) N N Preface This article explains the steps for Power Systems SAN multipath configuration using the SAN port virtualization feature NPIV. Using this virtualization feature, the SAN HBA port utilization can be increased and based on the bandwidth requirements it can reduce the overall HBAs required for SAN access. This document shows the configuration done using IBM AIX Operating System and shows the NPIV configuration part only. It does not explain the SAN switch configuration in detail or the details like VIOS installation, Virtual SCSI mapping, etc. IBM Corporation. 2010, 2011. All rights reserved. Page 2 of 9

Table of Contents 1. Introduction... 4 2. SAN Multipath Configuration using NPIV... 4 2.1. Pre-requisites... 4 2.2. VIOS Configuration... 5 2.3. AIX LPAR Configuration... 7 2.4. Configuring the SAN Storage... 8 3. References... 9 IBM Corporation. 2010, 2011. All rights reserved. Page 3 of 9

1. Introduction N_Port ID Virtualization or NPIV is a Fibre Channel facility allowing multiple N_Port IDs to share a single physical N_Port. This allows multiple Fibre Channel initiators to occupy a single physical port, easing hardware requirements in Storage Area Network design. An NPIV-capable fibre channel HBA can have multiple N_Port IDs, each with a unique identity and world wide port name. On System p the Virtual I/O Server (VIOS) allows sharing of physical resources between LPARs including virtual SCSI, virtual Fibre Channel (NPIV) and virtual networking. NPIV allows System p logical partitions (LPARs) to have dedicated N_Port IDs, giving the OS a unique identity to the SAN, just as if it had a dedicated physical HBA(s). For availing maximum redundancy for the SAN paths of the LPARs, the paths has to go through two different HBAs and those HBAs should be assigned to two different VIO Servers. The below given diagram (Figure 1.1) shows the logical representation of the configuration explained in this document. Figure 1.1 2. SAN Multipath Configuration using NPIV This section explains the pre-requisites and the steps involved in configuring the multipath SAN connectivity for the Power Systems using NPIV. 2.1. Pre-requisites This lists only those pre-requisites for the AIX Operating System which are required for doing the Multipath NPIV configuration. IBM POWER6 or above processor based hardware IBM Corporation. 2010, 2011. All rights reserved. Page 4 of 9

8Gb PCIe Dual-port FC Adapter (Feature Code #5735) SAN switch supporting the NPIV feature VIOS version 2.1 or above AIX 5.3 Technology Level 09 or above AIX 6.1 Technology Level 02 or above 2.2. VIOS Configuration You have to create dual VIOS instances on the System p server to avail maximum redundancy for the paths. Each VIOS should own at least 1 number of 8Gb FC Adapter (FC#5735) and one Virtual FC Server Adapter to map it to the client LPAR. NOTE: Before doing any NPIV related configuration, the VIOS and AIX LPAR profile configuration has to be completed with all the other details required for running the VIOS and the AIX LPAR, like proper Virtual SCSI adapter and/or Virtual Ethernet Adapter mapping, physical SCSI/RAID adapter mapping to the VIOS and/or AIX LPAR etc. This is because, for doing a Virtual FC Server Adapter to Client Adapter mapping both the VIOS and AIX LPAR should exist in the system, NOT like Virtual SCSI adapter mapping where you have to provide only the slot numbers and it does not check for the existence of the client. 2.2.1. Creating Virtual FC Server Adapters Open the first VIOS s profile for editing and navigate to the Virtual Adapters tab, click on Actions > Create > Fibre Channel Adapter as shown in the figure 2.1 Figure 2.1 Input the Virtual FC Server Adapter slot number, select the client LPAR from the dropdown which will be connected to this server adapter and input its planned Virtual FC Client Adapter slot number as shown in figure 2.2 Figure 2.2 IBM Corporation. 2010, 2011. All rights reserved. Page 5 of 9

Same way, edit the profile of the second VIOS and add the Virtual FC Server Adapter details as shown in figure 2.3 Figure 2.3 If you have any more LPARs which you want to configure for NPIV, repeat the above mentioned steps with those LPAR details. 2.2.2. Mapping Virtual FC Server Adapters to the physical HBA ports Once the Virtual FC Server Adapters are created, it has to be mapped to the physical HBA ports for connecting it to the SAN fabric. This will connect the Virtual FC Client Adapters to the SAN fabric through Virtual FC Server Adapter and physical HBA port. For mapping the Virtual FC Server Adapters to the HBA port, the HBA port should be already connected to the SAN switch enabled with NPIV support. You can run the command lsnports in VIOS $ prompt to know whether the connected SAN switch is enabled for NPIV. If the value for the fabric parameter shows as 1 that means that HBA port is connected to a SAN switch supporting the NPIV feature. If the SAN switch does not support NPIV, this value will be 0 and if there is no SAN connectivity there won t be any output for the lsnports command. If the value for fabric is 1, we can go ahead mapping the Virtual FC adapter to those HBA ports. In the VIOS, the Virtual FC adapters will be shown as vfchost devices. We can run the lsdev vpd grep vfchost command to know which device represents the Virtual FC adapter on any specific slot. In our setup, it shows that the vfchost0 is the server adapter device which is on slot 15. If there is more number of Virtual FC adapters, we can easily identify the device name if you know the server adapter slot number. Now we need to map the device vfchost0 to the physical HBA port fcs0 using the vfcmap -vadapter vfchost0 -fcp fcs0 command. Once it is mapped, check the status of the mapping using the command lsmap -vadapter vfchost0 npiv. Please note that the status of the port is showing as NOT_LOGGED_IN, this is because the client configuration is not yet complete and hence it cannot login to the fabric. IBM Corporation. 2010, 2011. All rights reserved. Page 6 of 9

Repeat the above mentioned steps in the second VIOS LPAR also. If you have more client LPARs, repeat the steps for all those LPARs in both the VIOS LPARs. 2.3. AIX LPAR Configuration Now we need to create the Virtual FC Client Adapters in the AIX LPAR and configure necessary drivers in the AIX OS to use multipathing. 2.3.1. Creating Virtual FC Client Adapters Open the LPAR profile for the AIX LPAR for editing and add the Virtual FC Client Adapters by navigating to the Virtual Adapters tab and selecting Actions > Create > Fibre Channel Adapter as shown in the figure 2.4 Figure 2.4 Input the first Virtual FC Client Adapter slot number details as shown in the figure 2.5. You need to make sure that the slot numbers entered is exactly matching with the slot numbers entered while creating the Virtual FC Server Adapter in the first VIOS LPAR. This adapter will be the first path for the SAN access in the AIX LPAR. Figure 2.5 Create the second Virtual FC Client Adapter with the slot number details as shown the figure 2.6. Make sure the slot numbers match with the slot numbers we have entered in the second VIOS LPAR while creating the Virtual FC Server Adapter. IBM Corporation. 2010, 2011. All rights reserved. Page 7 of 9

Figure 2.6 2.3.2. Configure AIX Operating System to use multipath Now activate the AIX LPAR and install AIX in it, note that the minimum version required to support the NPIV feature is AIX 5.3 TL9 or AIX 6.1 TL2. Once the AIX installation is complete, depending on the SAN Storage box, you need to install the necessary subsystem driver and configure it. To attach the storage, the storage vendor will specify what multi-path code must be used with their storage, and there may be options. We generally recommend multi-path code compliant with the MPIO architecture provided the storage supports it. For ESS, SVC, Storwize, DS6000, DS8000, DS3950, DS4000 (most models) and DS5000, you can use SDDPCM which is compliant with the MPIO architecture. This requires a "host attachment" fileset which populates the ODM to recognize and configure the storage correctly, and the SDDPCM code itself. The URL where you can download the SDDPCM is given below in References section. If you are using IBM s XIV storage, native XIV support is included in many of the latest AIX releases and separate driver installation may not be necessary, separate XIV Host Attachment Kit is required for older version of AIX. Consult your storage administrator if you are in any doubt of the driver requirements in AIX. You can now check the status of the Virtual FC Server Adapter ports in both the VIOS to check whether the ports are successfully logged in to the SAN fabric. You can use the lsmap vadapter vfchost0 npiv command. You can see that the port is logged into the SAN fabric and it shows the client LPAR details too. 2.4. Configuring the SAN Storage Once the AIX LPAR is activated and the Virtual FC ports are showing as logged in, you can configure your SAN storage box to do the proper Zoning and Access Control by providing the WWNs shown in the AIX LPAR. Note that the Virtual FC Client Adapter holds the WWNs for the clients and Virtual FC Server Adapters are just a connecting adapter used for passing through the SAN fabric data to the physical HBA port. You can get the WWNs of the Virtual FC Client Adapter ports either from the AIX operating system or from the Hardware Management Console. From AIX, find out the Virtual FC Adapter which is mapped to the first and second VIOS by seeing the slot numbers from the command output lscfg vpl fcs*. Once you have found the device name for the specific adapter, get the WWN for that adapter by using the command lscfg -vpl fcs0 grep Address as shown below. IBM Corporation. 2010, 2011. All rights reserved. Page 8 of 9

For seeing the WWNs of the Virtual FC Client Adapter from the HMC, open the properties for the LPAR and navigate to the Virtual Adapters tab, then select the Virtual FC Client Adapter for which you want to see the WWN and click on properties. NOTE: When viewing the properties of the Virtual FC Client Adapter from the HMC, it will show two WWNs for each Virtual FC Client Adapter as shown in figure 2.7. The second WWN shown here is not used until there is a live migration activated on this LPAR through Live Partition Mobility. Figure 2.7 Similar way, find out the WWN for the second adapter which is mapped through the second VIOS. Create the SAN LUNs, do proper Zoning and map the LUNs to these two adapters. If you are planning to use Live Partition Mobility on this LPAR, the Zoning for SAN access has to be done using both the WWNs shown in the figure 2.7. When a live migration happens for this LPAR, the new migrated hardware will be accessing SAN storage using the second WWN of the Virtual FC Client Adapter, so you have to make sure the second WWN is also configured in Zoning and Access Control. Once it s done run the cfgmgr command in AIX LPAR and check whether the disks are getting detected. Once it is detected you can check the status of multipathing. If the storage is IBM DS series, use the SDD commands datapath query adapter, datapath query device, lsvpcfg etc. If the storage is XIV, you can check the status of multipathing using the lspath command. It will show the output as shown below. You can see that there are two separate paths for the disk hdisk2 which is through two separate virtual FC adapters. 3. References S/N Item Link 1 IBM TechDocs http://w3-03.ibm.com/support/techdocs/atsmastr.nsf/webindex/fq128819 2 SDDPCM for IBM Storage http://www-01.ibm.com/support/docview.wss?rs=540&context=st52g7&uid=ssg1s4000201&loc=en_us&cs=utf-8&lang=en IBM Corporation. 2010, 2011. All rights reserved. Page 9 of 9