Linux Storage Area Network Topics & Linux Kernel Developers Summit Report

Similar documents
Open Source for OSD. Dan Messinger

Open Source support for OSD

STORAGE CONSOLIDATION WITH IP STORAGE. David Dale, NetApp

STORAGE CONSOLIDATION WITH IP STORAGE. David Dale, NetApp

Learn Your Alphabet - SRIOV, NPIV, RoCE, iwarp to Pump Up Virtual Infrastructure Performance

Voltaire. Fast I/O for XEN using RDMA Technologies. The Grid Interconnect Company. April 2005 Yaron Haviv, Voltaire, CTO

OFED Storage Protocols

Evaluating the Impact of RDMA on Storage I/O over InfiniBand

Configuration Maximums VMware vsphere 5.0

InfiniBand Networked Flash Storage

Storage Protocol Offload for Virtualized Environments Session 301-F

iser as accelerator for Software Defined Storage Rahul Fiske, Subhojit Roy IBM (India)

Windows Host Utilities Installation and Setup Guide

Comparing Server I/O Consolidation Solutions: iscsi, InfiniBand and FCoE. Gilles Chekroun Errol Roberts

<Insert Picture Here> DIF, DIX and Linux Data Integrity

Overview. Implementing Fibre Channel SAN Boot with the Oracle ZFS Storage Appliance. January 2014 By Tom Hanvey; update by Peter Brouwer Version: 2.

System input-output, performance aspects March 2009 Guy Chesnot

ESX 5.5 ESX 5.1 ESX 5 ESX

Creating an agile infrastructure with Virtualized I/O

I/O Considerations for Server Blades, Backplanes, and the Datacenter

NAS for Server Virtualization Dennis Chapman Senior Technical Director NetApp

Overview. Implementing Fibre Channel SAN Boot with Oracle's Sun ZFS Storage Appliance. August By Tom Hanvey; update by Peter Brouwer

Red Hat Global File System

Configuring iscsi in a VMware ESX Server 3 Environment B E S T P R A C T I C E S

A SCSI Transport Layer Extension with Separate Data and Control Paths for Scalable Storage-Area-Network Architectures

BUILDING A BLOCK STORAGE APPLICATION ON OFED - CHALLENGES

VMware Infrastructure Update 1 for Dell PowerEdge Systems. Deployment Guide. support.dell.com

Xen Summit Spring 2007

ETERNUS Multipath Driver (Windows Version) Installation Information

Active System Manager Release 8.2 Compatibility Matrix

PAC532 iscsi and NAS in ESX Server 3.0. Andy Banta Senior Storage Engineer VMware

Cisco - Enabling High Performance Grids and Utility Computing

Unified Storage and FCoE

Multifunction Networking Adapters

Dell 1741M Converged Network Adapter FCoE Boot from SAN Guide

Performance Report: Multiprotocol Performance Test of VMware ESX 3.5 on NetApp Storage Systems

Jake Howering. Director, Product Management

DESIGN AND IMPLEMENTATION OF A SOFTWARE PROTOTYPE FOR STORAGE AREA NETWORK PROTOCOL EVALUATION

W H I T E P A P E R. What s New in VMware vsphere 4: Performance Enhancements

Application Acceleration Beyond Flash Storage

Configuration Maximums. Update 1 VMware vsphere 6.5 VMware ESXi 6.5 vcenter Server 6.5

iscsi Initiator Design and Implementation Experience Kalman Z. Meth Abstract 1. Introduction 1.0 SANs and iscsi

MAGiC Conference September, 2017

Dell EMC SAN Storage with Video Management Systems

Virtualization. Santa Clara Valley Chapter of the IEEE Communication Society June 20, 2007 Scott Lurndal, 3Leaf Systems

Configuring and Managing Virtual Storage

USING ISCSI MULTIPATHING IN THE SOLARIS 10 OPERATING SYSTEM

Mellanox Technologies Maximize Cluster Performance and Productivity. Gilad Shainer, October, 2007

VMware VMmark V1.1 Results

Traditional SAN environments allow block

QuickSpecs. HP Integrity Virtual Machines (Integrity VM) Overview. Currently shipping versions:

COSC6376 Cloud Computing Lecture 17: Storage Systems

Cisco Exam Questions & Answers

Vendor: IBM. Exam Code: Exam Name: Storage Sales V2. Version: DEMO

USING ISCSI AND VERITAS BACKUP EXEC 9.0 FOR WINDOWS SERVERS BENEFITS AND TEST CONFIGURATION

Unified Storage Networking. Dennis Martin President, Demartek

Open-FCoE Software Initiator

QuickSpecs. HP Integrity Virtual Machines (Integrity VM) Overview. Currently shipping versions:

THE ZADARA CLOUD. An overview of the Zadara Storage Cloud and VPSA Storage Array technology WHITE PAPER

CompTIA ISS-001. Intel Server Specialist Certification. Download Full Version :

Configuration Maximums VMware Infrastructure 3: ESX Server 3.5 Update 2, ESX Server 3i version 3.5 Update 2, VirtualCenter 2.

Fibre Channel and iscsi Configuration Guide for the Data ONTAP 8.0 Release Family

HP Cluster Interconnects: The Next 5 Years

W H I T E P A P E R. Comparison of Storage Protocol Performance in VMware vsphere 4

What is RDMA? An Introduction to Networking Acceleration Technologies

iscsi or iser? Asgeir Eiriksson CTO Chelsio Communications Inc

Storage and Storage Access

IBM A Assessment: PureFlex Technical Expert V1.

Windows Unified Host Utilities 7.0 Installation Guide

HP Supporting the HP ProLiant Storage Server Product Family.

CIS 21 Final Study Guide. Final covers ch. 1-20, except for 17. Need to know:

vsphere Storage Update 1 Modified 16 JAN 2018 VMware vsphere 6.5 VMware ESXi 6.5 vcenter Server 6.5

Big Data Processing Technologies. Chentao Wu Associate Professor Dept. of Computer Science and Engineering

Linux iscsi.org BoF. Current Status and Future of iscsi on the Linux platform

Technical White Paper iscsi Boot November 11, 2004

Roadmap for Challenging Times System Virtualiztion

SNIA Developers Conference - Growth of the iscsi RDMA (iser) Ecosystem

VMware Infrastructure Update 1 for Dell PowerEdge Systems. Deployment Guide. support.dell.com

Vmware VCP-310. VMware Certified Professional on VI3.

The Last Bottleneck: How Parallel I/O can improve application performance

Red Hat Enterprise 7 Beta File Systems

Dell with Oracle Database 11g R2 on RHEL 6.3 & Oracle Linux 6.3 with Compatible Kernel

Module 2 Storage Network Architecture

Red Hat Virtualization 4.1 Technical Presentation May Adapted for MSP RHUG Greg Scott

PCI Express x8 Single Port SFP+ 10 Gigabit Server Adapter (Intel 82599ES Based) Single-Port 10 Gigabit SFP+ Ethernet Server Adapters Provide Ultimate

Use of the Internet SCSI (iscsi) protocol

IBM IBM PureFlex Technical Expert V1.

Additional FCoE Resources

FCoE at 40Gbps with FC-BB-6

Informatix Solutions INFINIBAND OVERVIEW. - Informatix Solutions, Page 1 Version 1.0

Next Generation Computing Architectures for Cloud Scale Applications

Configuration Maximums

The Last Bottleneck: How Parallel I/O can attenuate Amdahl's Law

Interoperability of Bloombase StoreSafe and Huawei OceanStor V3-Series for Transparent Storage Encryption

QuickSpecs. HP Integrity Virtual Machines (Integrity VM) Overview. Retired. Currently shipping versions:

Oracle VM. Getting Started Guide for Release 3.2

Software-defined Shared Application Acceleration

Open Source Storage. Ric Wheeler Architect & Senior Manager April 30, 2012

The File Systems Evolution. Christian Bandulet, Sun Microsystems

Fibre Channel. Roadmap to your SAN Future! HP Technology Forum & Expo June, 2007, Las Vegas Skip Jones, FCIA Chair, QLogic

Transcription:

Linux Storage Area Network Topics & Linux Kernel Developers Summit Report FUJITA Tomonori

Storage Area Network topics

SCSI Client Server Model Initiator device (client) Sending SCSI requests Target device (Server) Performing SCSI requests and sending responses Transport Connecting initiators with targets From t10.org

What is Storage Area Network (SAN)? hosts and storage devices are connected via network Linux implementation Block Commands (disk) Stream Commands (tape) Multi-media Commands (cd, DVD) Upper level Drivers (ULD) SCSI Primary Command set (for all device types) Mid Layer Parallel SCSI Fibre Channel RDMA iscsi (TCP) Serial Attached SCSI Low level Drivers (LLD)

What is the point in using virtual machines with SAN? Shared external storage with VMM Enhancing manageability and availability Booting a host from a LUN on a SAN VMM runs on a diskless box (maybe blade) Virtual machine cluster with SAN and multiple hosts failover

SAN storage systems (target devices) are expensive!

Can we convert a Linux box to a SAN storage system?

Target mode support Building a SCSI target device by using a Linux box to provide device services Interpreting SCSI requests and sending back the results to initiators Why do we want it? Poor man s SAN target devices For debugging initiator drivers

How to design target mode? Monolithic target mode design is wrong e.g. iscsi target software, FC target software Stop reinventing the wheel Executing SCSI commands SCSI state machine logical unit management

Target infrastructure and target mode drivers Transport Protocols Hardware Interface Target drivers & kernel subsystems FCP FC HBA Target driver NIC TCP stack Target driver iscsi iscsi HBA Target driver Target Infrastructure IB/RNIC IB stack Target driver SRP RNIC IB TCA RNIC stack IB stack Target driver Target driver

SCSI target infrastructure (aka tgt) design tgt core is integrated with SCSI Mid layer and transport classes Target mode driver Target mode driver Transport libraries Management tool Unix socket User space tgtadm tgt daemon Main process Kernel space Target mode driver Target mode driver Target mode driver tgt core SCSI Mid layer Transport class Disk I/Os

Device type code Tgtcan provide a file as disk provide an iso image file as cdrom Tgt will provide a file as tape (Virtual Tape library) provide a file as object storage

Supported tgt drivers iscsi software driver (user space) In RHEL 5.1 beta, opensuse and debian (sid) Trying to beat iscsi enterprise target (aka IET) iscsi Extension for RDMA driver (user space) The code exists, but I ve not merged it yet Qla2xxx FC driver Under Qlogic review IBM System p Virtual target driver SRP target devices (one VM provides the rest device service, like Xen blkback dirver)

bsg (block layer scsi generic driver or sg v4) Sg v3 (drivers/scsi/sg.c) shortcomings Can handle only SCSI request/response for only scsi devices We need transport level commands for management We need SCSI task management commands Can t commands with bidirectional data Can t handle commands with task tag or attribute

bsg (cont.) Generic request/response protocol interface You can create a bsg device for any object in kernel (bsg can be bind with request_queue) You can handle any protocol (define your own request and response handlers) User space can send requests and receive response via read/write system calls

bsg: SAN management SAS Management Protocol (SMP) Talking with SAS devices and expanders via HBA bsg devices are bind with every SAS object bsg provides an unified interface Stop vendors from inventing various interfaces aic94xx and mptsas are supported User space Kernel space SMP utility SMP interface SAS transport class libsas aic94xx SAS HBAs mptsas SAS HBAs

SMP example REPORT_MANUFACTURE Ask who you are # smp_test /sys/class/bsg/expander-2:0 SAS-1.1 format: 0 vendor identification: LSILOGIC product identification: SASx12 A.0 product revision level:

bsg: bidirectional data transfer (bidi) Object Storage Devices (aka OSD) variable-length objects and object ids bidi support is a must Bidi support status Block layer is in upstream (needed for SMP) Scsi mid layer is not yet (hopefully 2.6.25)

scsi data accessors Insulating LLDs from data transfer information We plan to make lots of changes to scsi_cmnd structure support sg chaining and bidirectional data transfer LLDs directly accessed to the values in scsi_cmnd We rewrite LLDs to access scsi_cmnd via new accessors

scsi data accessors (cont.) Too simple sg setup examples How a LLD tell addresses for I/Os for the HBA Old way struct scsi_cmnd *sc struct scatterlist *sg = sc->request_buffer; for(i = 0; i < nseg; i++) { paddr = sg_dma_address(sg[i]); New way stcuct scsi_cmnd *sc struct scatterlist *sg; scsi_for_each_sg(sc, sg, nseg,i) { physaddr = sg_dma_address(sg);

scatter gather chaining SCSI-ml couldn t handle Large data transfer scsi-ml pools 8, 16, 32, 128 sg entries (the sg size is 32 bytes on x86_64) People complains about scsi memory consumption so we can t have large sg entries scsi_cmnd struct has a point to sg entries

scatter gather chaining (cont.) sg chaining The last entry in a sg list points to the next sg list sg entries aren t continuous any more! scsi_cmnd structure SG entries The maximum of the entries are 7 + 7 + 8. SG entries SG entries

How didi scsi data accessors help sg chaining? Before sg chaining #define scsi_for_each_sg(sc, sg, nseg, i) for(i = 0, sg = scsi_sglist(sc); i < nseg, i++, sg++) sg entries must be continuous We changed it after sg chaining #define scsi_for_each_sg(sc, sg, nseg, i) for(i = 0, sg = scsi_sglist(sc); i < nseg, i++, sg = sg_next(sg)) sg_next macro takes care of discontinuous sg entries LLDs can support sg chaining magically without modifications

Future SAN technologies

Fibre Channel over Ethernet (FCoE) Putting a FC frame into an Ethernet frame iscsi? Why do we need TCP overhead in data center? TOE/RDMA? Who needs complicated hardware? Intel plans to release software FCoE initiator driver in December Should be easy to add FCoE software target driver to tgt FCoE might be a good solution for VMs

Object Storage Device (OSD) New SCSI device type No sector, initiators can access to variable-length objects with 64-bit id Who needs OSD? pnfs (parallel NFS) Parallel file systems (lustre, IBM storage tank, etc) prefer object-based storage Local file systems for OSD?

Linux Kernel Developers Summit

Agenda day1 Distributor kernel maintainer panel Mini-summit reports The greater kernel ecosystem Documentation Kernel quality Hardware I386/x86_64 merger

Agenda day2 Customer panel Real-time and scheduler Scalability Memory management Containers Developer relations Some of Andrew s process questions Future kernel summits