NetApp Solutions for Hadoop Reference Architecture

Similar documents
Architecting Storage for Semiconductor Design: Manufacturing Preparation

Dell PowerEdge R720xd with PERC H710P: A Balanced Configuration for Microsoft Exchange 2010 Solutions

Assessing performance in HP LeftHand SANs

Hadoop File System S L I D E S M O D I F I E D F R O M P R E S E N T A T I O N B Y B. R A M A M U R T H Y 11/15/2017

CS370 Operating Systems

EMC Business Continuity for Microsoft Applications

An Introduction to GPFS

SGI Hadoop Based on Intel Xeon Processor E5 Family. Getting Started Guide

Microsoft Office SharePoint Server 2007

EsgynDB Enterprise 2.0 Platform Reference Architecture

Fusion iomemory PCIe Solutions from SanDisk and Sqrll make Accumulo Hypersonic

IBM System Storage DS5020 Express

Dell Fluid Data solutions. Powerful self-optimized enterprise storage. Dell Compellent Storage Center: Designed for business results

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

JMR ELECTRONICS INC. WHITE PAPER

TITLE: PRE-REQUISITE THEORY. 1. Introduction to Hadoop. 2. Cluster. Implement sort algorithm and run it using HADOOP

NetApp High-Performance Storage Solution for Lustre

Configuring a Single Oracle ZFS Storage Appliance into an InfiniBand Fabric with Multiple Oracle Exadata Machines

An Oracle Technical White Paper October Sizing Guide for Single Click Configurations of Oracle s MySQL on Sun Fire x86 Servers

DELL Reference Configuration Microsoft SQL Server 2008 Fast Track Data Warehouse

Building Backup-to-Disk and Disaster Recovery Solutions with the ReadyDATA 5200

Test Report: Digital Rapids Transcode Manager Application with NetApp Media Content Management Solution

Video Surveillance Storage and Verint Nextiva NetApp Video Surveillance Storage Solution

Cloudera Exam CCA-410 Cloudera Certified Administrator for Apache Hadoop (CCAH) Version: 7.5 [ Total Questions: 97 ]

Distributed Filesystem

Dell PowerVault MD Family. Modular Storage. The Dell PowerVault MD Storage Family

White Paper. EonStor GS Family Best Practices Guide. Version: 1.1 Updated: Apr., 2018

A Thorough Introduction to 64-Bit Aggregates

Deploying and Managing Dell Big Data Clusters with StackIQ Cluster Manager

IBM System Storage DCS3700

EMC Virtual Infrastructure for Microsoft Exchange 2010 Enabled by EMC Symmetrix VMAX, VMware vsphere 4, and Replication Manager

Veritas NetBackup on Cisco UCS S3260 Storage Server

QuickSpecs. Models SATA RAID Controller HP 6-Port SATA RAID Controller B21. HP 6-Port SATA RAID Controller. Overview.

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

Extremely Fast Distributed Storage for Cloud Service Providers

CA485 Ray Walshe Google File System

Storage Consolidation with the Dell PowerVault MD3000i iscsi Storage

Vendor must indicate at what level its proposed solution will meet the College s requirements as delineated in the referenced sections of the RFP:

Overview of HP tiered solutions program for Microsoft Exchange Server 2010

Accelerating Big Data: Using SanDisk SSDs for Apache HBase Workloads

Dell PowerVault MD Family. Modular storage. The Dell PowerVault MD storage family

NEC M100 Frequently Asked Questions September, 2011

High-Performance Lustre with Maximum Data Assurance

HCI: Hyper-Converged Infrastructure

Surveillance Dell EMC Storage with Bosch Video Recording Manager

BIG DATA AND HADOOP ON THE ZFS STORAGE APPLIANCE

The advantages of architecting an open iscsi SAN

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

Distributed File Systems II

Microsoft SQL Server in a VMware Environment on Dell PowerEdge R810 Servers and Dell EqualLogic Storage

Video Surveillance EMC Storage with LENSEC Perspective VMS

Lenovo RAID Introduction Reference Information

A GPFS Primer October 2005

SurFS Product Description

How Smarter Systems Deliver Smarter Economics and Optimized Business Continuity

DRIVESCALE-HDP REFERENCE ARCHITECTURE

MI-PDB, MIE-PDB: Advanced Database Systems

Virtual Security Server

RAIDIX Data Storage Solution. Clustered Data Storage Based on the RAIDIX Software and GPFS File System

IBM Scale Out Network Attached Storage (SONAS) using the Acuo Universal Clinical Platform

Can Enterprise Storage Fix Hadoop? PRESENTATION TITLE GOES HERE John Webster Senior Partner Evaluator Group

White Paper. Low Cost High Availability Clustering for the Enterprise. Jointly published by Winchester Systems Inc. and Red Hat Inc.

DRIVESCALE-MAPR Reference Architecture

EMC CLARiiON CX3-40. Reference Architecture. Enterprise Solutions for Microsoft Exchange 2007

Target Environments The Smart Array 6i Controller offers superior investment protection to the following environments: Non-RAID

IBM System Storage DS4800

Storageflex HA3969 High-Density Storage: Key Design Features and Hybrid Connectivity Benefits. White Paper

IBM TS7700 grid solutions for business continuity

High Availability and Disaster Recovery features in Microsoft Exchange Server 2007 SP1

Cisco Tetration Analytics Platform: A Dive into Blazing Fast Deep Storage

International Journal of Advance Engineering and Research Development. A Study: Hadoop Framework

A Thorough Introduction to 64-Bit Aggregates

Software Defined Storage at the Speed of Flash. PRESENTATION TITLE GOES HERE Carlos Carrero Rajagopal Vaideeswaran Symantec

Virtual Disaster Recovery

HP ProLiant m300 1P C2750 CPU 32GB Configure-to-order Server Cartridge B21

Dell EMC PowerEdge and BlueData

Performance Report: Multiprotocol Performance Test of VMware ESX 3.5 on NetApp Storage Systems

Reference Architecture Microsoft Exchange 2013 on Dell PowerEdge R730xd 2500 Mailboxes

Surveillance Dell EMC Storage with Verint Nextiva

Dell Reference Configuration for Hortonworks Data Platform 2.4

Veeam Availability Solution for Cisco UCS: Designed for Virtualized Environments. Solution Overview Cisco Public

Dell Reference Configuration for Large Oracle Database Deployments on Dell EqualLogic Storage

QuickSpecs. Models. Overview

Microsoft Exchange Server 2010 workload optimization on the new IBM PureFlex System

Introduction to NetApp E-Series E2700 with SANtricity 11.10

EMC CLARiiON CX3-40. Reference Architecture. Enterprise Solutions for Microsoft Exchange Enabled by MirrorView/S

RAIDIX Data Storage System. Entry-Level Data Storage System for Video Surveillance (up to 200 cameras)

NetApp SolidFire and Pure Storage Architectural Comparison A SOLIDFIRE COMPETITIVE COMPARISON

Using the ThinkServer SA120 Disk Array

Hitachi Adaptable Modular Storage and Workgroup Modular Storage

Virtualization And High Availability. Howard Chow Microsoft MVP

EMC Business Continuity for Microsoft SharePoint Server (MOSS 2007)

HP Supporting the HP ProLiant Storage Server Product Family.

Real-time Protection for Microsoft Hyper-V

Hitachi Adaptable Modular Storage and Hitachi Workgroup Modular Storage

RAIDIX Data Storage Solution. Highly Scalable Data Storage System for Postproduction Studios and Broadcasting Corporations

EMC Backup and Recovery for Microsoft SQL Server

CS370 Operating Systems

DATA PROTECTION IN A ROBO ENVIRONMENT

IBM InfoSphere Streams v4.0 Performance Best Practices

Transcription:

White Paper NetApp Solutions for Hadoop Reference Architecture Gus Horn, Iyer Venkatesan, NetApp April 2014 WP-7196 Abstract Today s businesses need to store, control, and analyze the unprecedented complexity, speed, or volume of data they are capturing. Hadoop has gained popularity by handling these large and diverse data types. However, there are technical challenges that enterprises face in deploying Hadoop, specifically in the areas of cluster availability, operations, and scale. NetApp has developed reference architectures with leading Hadoop vendors to deliver a solution that overcomes some of these challenges so that business can ingest, store, and manage this big data with higher reliability and scalability, and with less time spent on operations and maintenance. This white paper discusses a flexible, validated, enterprise-class Hadoop architecture based on NetApp storage in a managed or external direct-attached storage (DAS) configuration with any compatible distribution of Apache Hadoop. Included are guidelines and best practices on how to build and configure modular Hadoop systems, including recommendations on servers, file systems, and networking topology. This enables groups or organizations to gain insights for competitive advantage, in-depth customer knowledge, and many other business benefits.

TABLE OF CONTENTS 1 Introduction... 4 1.1 Benefits...4 1.2 Brief Hadoop Overview...4 1.3 Basic Hadoop Architecture...5 2 Hadoop Challenges... 5 2.1 HDFS Replication Trade-Offs...6 2.2 When Drives Fail...6 3 NetApp Solutions for Hadoop Architecture... 7 3.1 High-Level Architecture...7 3.2 Technical Advantages of the NetApp Solutions for Hadoop...7 3.3 Validated Base Configuration...7 3.4 Other Supported Configurations...8 4 Solution Architecture Details... 8 4.1 Network Architecture...9 4.2 Storage Architecture... 10 4.3 NameNode Metadata Protection... 11 4.4 Rack Awareness Implementation... 13 5 Key Components of Validated Configuration... 13 6 Other Supported Features, Products, and Protocols... 14 7 Conclusion... 14 Additional Information... 15 Version History... 15 LIST OF TABLES Table 1) Recommended products for the validated reference architecture.... 13 Table 2) Components supported by NetApp and its partners.... 14 LIST OF FIGURES Figure 1) Basic Hadoop architecture....5 Figure 2) Architecture diagram for NetApp solutions for Hadoop....7 Figure 3) Network architecture.... 10 Figure 4) E-Series configured with four hosts per array.... 11 2 NetApp Solutions for Hadoop Reference Architecture

Figure 5) E-Series configured with eight hosts per array.... 11 Figure 6) NameNode backup configuration using NFS.... 12 Figure 7) NameNode HA configuration using NFS storage.... 12 3 NetApp Solutions for Hadoop Reference Architecture

1 Introduction At present, the term big data is well known and is having a significant impact on businesses that need to store, manage, and analyze the complex and large amounts of data that they are capturing. Hadoop and its growing ecosystem of products have enabled many businesses to handle this volume and these diverse types of data so they can start gaining valuable insights from this data. However, Hadoop poses some challenges in enterprise data centers, centering on operations, availability, and implementation. Validated reference architectures for Hadoop are an effective way to overcome some of these challenges in implementing Hadoop. NetApp Solutions for Hadoop validate NetApp storage with the leading Hadoop distributions in a flexible and open environment. Recommendations for servers and networking are also included so a fully configured cluster can be built. Asgroups begin to move Hadoop pilots or proofs of concept into production; they typically expect a certain degree of scalability, manageability, and availability as well as consistent, deterministic performance. The NetApp reference architecture is designed for HA scale out and will provide SLAs with deterministic performance. In short it is an Open enterprise grade scale out solution. 1.1 Benefits Key benefits of the NetApp solutions for Hadoop are: Enterprise-class implementation of Hadoop with lower cluster downtime, higher data availability, and linear scalability Fast time to value with the validated, presized configurations that are ready to deploy, thereby reducing risks that are traditionally associated with Hadoop High storage efficiency and lower operational expenses for Hadoop Open solution built with NetApp and best of breed products from partners, providing choice without vendor lock-in or performance compromise NetApp has partnered with major Hadoop distribution providers and other analytic platform vendors to deliver reliable and efficient storage for Hadoop solutions. These validated solutions when implemented in a Hadoop cluster provide the capabilities for ingesting, storing, and managing large datasets with high efficiency, reliability and performance. 1.2 Brief Hadoop Overview Readers familiar with Hadoop can skip this section and can either scroll to section 1.3, Basic Hadoop Architecture, or section 2, Hadoop Challenges. Hadoop is an open source analytical framework and an ecosystem of products and technologies that contain, among other things, databases, scripting languages, and management tools. Hadoop is designed to run on a large number of parallel machines. When you want to load all of your organization s data into Hadoop, the software makes three copies of the data, then breaks that data into pieces and spreads it across different servers that are set up to run in parallel. There is no one place where you can go to manage all of your data; Hadoop NameNode keeps track of where the data resides. And because there are multiple copy stores, data stored on a server that goes offline or dies can be automatically replicated from a known good copy. The two main components of Hadoop are MapReduce, which is the framework for parallel processing. The other one is Hadoop Distributed File System (HDFS), which is the distributed file system that provides petabyte-size storage and data management. 4 NetApp Solutions for Hadoop Reference Architecture

1.3 Basic Hadoop Architecture The basic Hadoop architecture is built around commodity servers and internal DAS storage and networking. Following are the basic requirements of this architecture. A conventional Hadoop cluster consists of a few basic components: NameNode/active NameNode. This manages the HDFS namespace. It runs all of the services needed to manage the HDFS data storage and MapReduce tasks in the cluster, including data and task distribution and tracking. This is sometimes called the NameNode or active NameNode. Checkpoint node. It is a secondary NameNode that manages the on-disk representation of the NameNode metadata. In active/standby HA, this node runs a second NameNode process, usually called the standby NameNode process. In quorum-based HA mode, this node runs the second journal node. JobTracker node. It manages all of the jobs submitted to the Hadoop cluster and facilitates job and task scheduling. DataNodes. These are slave nodes that provide both TaskTracker functionality and DataNode functionality. These nodes perform all of the real processing work done by the cluster. As DataNodes, they store all of the data blocks that make up the file system, and they serve I/O requests. As TaskTrackers, they perform the job tasks assigned to them by the JobTracker. Figure 1) Basic Hadoop architecture. 2 Hadoop Challenges Hadoop was incubated and developed at Internet companies including Google and Yahoo. Their goal was to practically analyze billions of files and petabytes of data using thousands of low-cost servers. They achieved this in an environment where development and operations coexist. As Hadoop gains traction in enterprise data centers, operational requirements change. A cluster with a few hundred nodes suffices for most businesses; enterprise customers do not have active developers working on Hadoop. Enterprises are driven by very strict SLAs. Any solution put into such environments must meet enterprise operational requirements and data center standards and SLAs. Following are some common challenges that are faced while implementing Hadoop in an enterprise: Operational 5 NetApp Solutions for Hadoop Reference Architecture

Limited flexibility when scaling, cannot independently add servers or storage (for example: when adding servers, storage has to be added as well). Not highly available. Jobs get interrupted during failures, and recovering these jobs from drive failures can be time consuming. Might not meet enterprise-level SLAs. Efficiency Three copies of data (also called triple mirroring). Network congestion and overdrain of system resources. Poor job efficiency due to storage fragmentation over a period of time. Management Requires skilled resources. Flexibility of changing architectures. Data center footprint. 2.1 HDFS Replication Trade-Offs Although HDFS is distributed, feature rich, and flexible, there are both throughput and operational costs associated with server-based replication for ingest, redistribution of data following recovery of a failed DataNode, and capacity utilization. A standard Hadoop install implements at least three copies of every piece of data stored. Costs of triple replication include: For each usable terabyte (TB) of data, 3TB of raw capacity are required. Server-based triple replication creates a significant load on the servers themselves. At the beginning of an ingest task, before any data blocks are actually written to a DataNode, the NameNode creates a list of DataNodes where all three of the first block replicas are written, forming a pipeline and allocating blocks to those nodes. That DataNode forwards the same block to the next node in the pipeline and the preceding procedure is repeated. Finally, this second DataNode sends the block to the third DataNode, and the write to storage follows the same I/O path as the former writes. This results in a significant load on server resources. Server-based replication also creates a significant load on the network. For ingest, three network trips are required to accomplish a single block write. All writes compete for and consume the same network, CPU, memory, and storage resources allocated to process the Hadoop analysis tasks. The NetApp solution can provide better reliability with 1/3 reduction in replication that is the NetApp architecture needs only two copies of data with our strictly shared nothing architecture. Anything done to reduce the consumption of server and network resources during the loading of data increases cluster efficiency. 2.2 When Drives Fail The most fault prone component in the Hadoop architecture are the disk drives, it is simply a matter of time when a drive will fail, and the larger the cluster, the higher the possibility that a drive failure will occur. If one or more drives fail, all tasks accessing data on that drive will fail and be reassigned to other DataNodes in the cluster. This results in a severe degradation of completion time for running jobs. In addition, the drive must be physically replaced, mounted, partitioned, formatted, and reloaded with data from the nearby copies before HDFS can make full use of that drive. Repopulating the new disk with data can be accomplished by using the Hadoop balancer, which redistributes HDFS data across all DataNodes by moving blocks from overutilized to underutilized DataNodes. For every full 3TB SATA disk that fails, several hours are needed for rebalancing the distribution of blocks following replacement of that disk, depending on the tolerance for network resource consumption in the 6 NetApp Solutions for Hadoop Reference Architecture

cluster. The trade-off is between rebalancing time, network utilization, and disk I/O bandwidth utilization. Remember that higher resource utilization affects MapReduce job performance, especially during data ingest operations. In some cases, rebalancing might also require some manual movement of blocks between disk drives. 3 NetApp Solutions for Hadoop Architecture 3.1 High-Level Architecture Figure 2 is the general architecture diagram for NetApp solutions for Hadoop reference designs. Figure 2) Architecture diagram for NetApp solutions for Hadoop. 3.2 Technical Advantages of the NetApp Solutions for Hadoop Lower cluster downtime with transparent RAID operations and faster rebuild of failed media Less storage needed with Hadoop replication count of two Enhanced robustness of HDFS through reliable NFS server-based metadata protection Disk drives can be replaced while the DataNode is running; this operation cannot be done with the internal DAS-based DataNode configurations Performance for the next generation of DataNodes (SMP) and networks (10GbE) Highly reliable and manageable Hadoop platform Server, storage, and networking hardware preconfigured and validated with Apache-compatible Hadoop distributions 3.3 Validated Base Configuration The NetApp solution for Hadoop has a validated base reference architecture that constitutes the following hardware, software, and networking. This recommended configuration has been tested, validated, and optimized for running a Hadoop cluster. 7 NetApp Solutions for Hadoop Reference Architecture

Servers Supports all of the Intel Xeon E5-2400 series processors Up to two 2.4GHz (6-core) or 2.3GHz (8-core) processors Memory: 12 DIMMs (up to 192GB), up to 1600Mhz Hard disks: 2 x 3.5" HS HDDs Network: 2x1Gb/sec Ethernet ports, 10GbE network port LSI 9207-8e SAS HBA or compatible ISCSI/FC HBA needed SFF-8088 external mini-sas cable (for maximum signal integrity, cable should not exceed 5m) Storage NetApp E5460 storage array 1 DE6600 4U chassis with rail kit and power cords 2 E-5460 series controllers, each configured with dual SAS ports, and SAS disk drives 60 3TB, 7.2K-RPM, 3.5" near-line NL-SAS disks NetApp Engenio operating system (EOS CFW 10.80 or later release) Base SANtricity Storage Manager Turbo feature enabled Optional: NetApp FAS2220 NFS storage system NetApp FAS2220 is recommended if customers require high-availability (HA) protection of the NameNode or need simplification in managing Hadoop clusters. Network 10GbE nonblocking network switch GbE network switch Software Red Hat Enterprise Linux 6.2 or later One of the following Hadoop distributions: Cloudera Distribution 4.1.3 or later; this includes Cloudera Manager and Cloudera Enterprise Hortonworks Data Platform HDP 1.3 or later NetApp SANtricity 10.84 or later storage management (bundled with NetApp E-Series storage arrays) 3.4 Other Supported Configurations Apart from the validated based configuration, this solution offers flexibility in choosing other models and versions of each component. You might be able to choose a lower or higher version of storage arrays based on your requirements. You might also consider a high-availability pair of network switches. Some users prefer to go after a bonded GbE NIC pair instead of a 10GbE network. Storage to DataNode connectivity also offers multiple options from direct connect SAS, FCP, or ISCSI. Section 6, Other Supported Features, Products, and Protocols, of this report provides a detailed list of such available options. 4 Solution Architecture Details In Figure 2, there are two pieces of NetApp storage in the reference design. The first is the NetApp E- Series storage array, which provides the data storage services with one E-Series dedicated to four 8 NetApp Solutions for Hadoop Reference Architecture

Hadoop DataNodes. The second optional piece is the FAS2220, which provides enterprise-grade protection of the NameNode metadata, if that is needed for the cluster. Each DataNode has its own dedicated, nonshared set of disks. This exactly follows the way in which Hadoop DataNodes configure capacity when these drives are physically colocated in the chassis, as in the internal JBOD configurations. In practical terms, Linux is not able to distinguish between a set of logical unit numbers (LUNs) presented by the LSI 1068 controller chip on the HBA and a set of disks presented by either an embedded SAS or a SATA controller. There is no logical difference. However, there are differences in performance and reliability. This set of disks is configured as two blocks, each using one of two parity-based protection schemes. Eight volume groups are configured in the E-Series, and they are configured so that each DataNode sees only its set of two private blocks. This design improves just a bunch of disks (JBOD) DAS media for four DataNodes by offering better performance, reliability, data availability, uptime, and manageability. Each block is a RAID-protected volume group that contains a single virtual logical volume, which can be created and exported to a given DataNode as a LUN. These LUNs appear as physical disks in Red Hat Enterprise Linux (RHEL). They can be partitioned, and file systems can be created and mounted to store HDFS blocks. (HDFS blocks can be 64MB, 128MB, 256MB, or even 512MB in size.) 4.1 Network Architecture The network architecture of a Hadoop cluster is critical. With a default replication count of three, more than two-thirds of all the I/O traffic must traverse the network during ingest, and perhaps a third of all MapReduce I/O during processing runs could originate from other nodes. The NetApp E-Series storage modules provide a level of performance that is significantly better than that provided by JBOD SATA. The storage array offers a performance platform based on four 6Gb/sec SAS ports, supported by proven caching and best-in-class hardware RAID. The solution uses two network backplanes. The core of this solution is the HDFS network, which is a 10GbE network backplane served by an extremely high-performance network 10/40GbE switch solution for a top-of-rack (ToR) switch. The other network is a GbE network, which caters to the administration network and also connects the NameNode, JobTracker, and NetApp FAS storage systems (if a FAS is used for NameNode HA). On the 10GbE network, all components should be configured for jumbo frames. This requirement also applies to the switch ports to which the components are connected. Any ports on the 10GbE switches that are configured for GbE connections (such as FAS NFS and uplinks) should not have jumbo frames enabled. 9 NetApp Solutions for Hadoop Reference Architecture

Figure 3) Network architecture. Recommended Network Topology Data network (10GbE) Private interconnect between all DataNodes 10GbE nonblocking switch 40GbE uplink between the racks Dedicated nonroutable VLAN Primary purpose: Data ingest and movement within cluster Use dedicated separate VLAN for iscsi network, if opted Management network (GbE) System administration network Publically routable subnet NameNode and JobTracker also use GbE All nodes have a public GbE interface for administration and data transfer purpose E-Series storage systems also need two GbE ports for management purpose only 4.2 Storage Architecture Each NetApp E-Series storage array is configured as eight independent RAID groups of seven disks that can be set up in a RAID 5 (2 x 6+1) configuration. This consumes 56 (8 x 7) disks. The remaining four disks are global hot spares. If customers deploy all of their files with a replication count of two, then using a single-parity drive over six spindles provides a good balance between storage space utilization and 10 NetApp Solutions for Hadoop Reference Architecture

RAID protection. As a result, there are constantly two forms of protection when the E-Series storage arrays are running with files set to a replication count of two. Users have an option to select two available fans in ratios based on their requirements. The more common one is having one E-Series storage server with 4 DataNodes, as shown in Figure 3. For workloads where lower storage density is enough, a lighter configuration of single E-Series storage serving 8 DataNodes may be considered. This is depicted in Figure 5. Figure 4) E-Series configured with four hosts per array. Figure 5) E-Series configured with eight hosts per array. 4.3 NameNode Metadata Protection Prior to Hadoop 2.0, the NameNode was a single point of failure (SPOF) in an HDFS cluster. Each cluster had a single NameNode, and if that machine or process became unavailable, the cluster as a whole would be unavailable until the NameNode was either restarted or brought up on a separate machine. There are two ways to mitigate such a risk of NameNode failure and the data loss: backup/restore if using Hadoop 1.x or using the HDFS high-availability feature in Hadoop 2.0 or later versions. Option 1: Backup/Restore Method If you are planning to go with the traditional Hadoop method, then the NetApp solution offers a unique feature of NameNode metadata protection. This protection is offered by introducing a NetApp FAS NFS storage system. The NameNode is configured to keep a copy of its HDFS metadata (FSImage) in the NFS mounted storage. The same NFS mounted storage is also made available to the secondary NameNode. In the case of a NameNode failure, the critical metadata is still safe on the NetApp FAS storage system. At this point, we can recover the metadata and restart services on the secondary NameNode to get the HDFS back and running promptly. 11 NetApp Solutions for Hadoop Reference Architecture

Figure 6) NameNode backup configuration using NFS. Option 2: HDFS High Availability Versions 2.0 and later of HDFS offer a high-availability NameNode feature to mitigate NameNode failures. The HDFS high-availability feature addresses the preceding problems by providing the option of running two redundant NameNodes in the same cluster in an active-passive configuration with a hot standby. This allows a fast failover to a new NameNode in the case that a machine crashes or a graceful administrator-initiated failover for the purpose of planned maintenance. In order for the standby node to keep its state synchronized with the active node, the current implementation requires that the two nodes both have access to a directory on a shared storage device. This is facilitated by a NetApp FAS NFS storage device in the solution, as stated in the following section. Figure 7) NameNode HA configuration using NFS storage. For more details about this configuration, visit http://hadoop.apache.org/docs/current2/hadoopyarn/hadoop-yarn-site/hdfshighavailabilitywithnfs.html. For either option, FAS is recommended because it helps manage the Hadoop cluster by storing boot images and binaries simplifying administration tasks. 12 NetApp Solutions for Hadoop Reference Architecture

4.4 Rack Awareness Implementation Implementing rack awareness in our Hadoop solution offers important benefits. Using this feature allows us to configure Hadoop and know the topology of the network: Rack awareness enables Hadoop to maximize network utilization by favoring block transfers within a rack over transfers between racks. The JobTracker is able to optimize MapReduce job performance by assigning tasks to nodes that are closer to their data in terms of network topology. By default, during HDFS writes, the NameNode assigns the second and third block replicas to nodes in a different rack from the first replica. This provides data protection even against rack failure; however, this is possible only if Hadoop has been configured with knowledge of its rack configuration. By utilizing the rack awareness feature, we can also make sure that the second copy of each block always gets stored on a different E-Series-based DataNode. This offers tolerance to storage array and DataNode failures. Care must be taken to make sure that for each HDFS block, the corresponding replica is stored on another E-Series controller. In multirack clusters, providing the actual rack location of each DataNode will accomplish this. If all the DataNodes are located in the same rack, rack mapping must be extended to reflect the E-Series enclosure or E-Series controller used for DataNode storage. Basically, use the HDFS rack awareness configuration to prevent data loss in the unlikely event of storage controller failure. 5 Key Components of Validated Configuration Table 1 summarizes the recommended products for the validated reference architecture. Table 1) Recommended products for the validated reference architecture. Component Product or Solution Details Storage NetApp E-Series 5460 storage array FAS2220 (if using for NameNode protection) E5460: 60 3TB drives for 180TB capacity 4 hot spares 8 volume groups; 8 volumes RAID 5 volume groups, 7 disks each Servers Intel Xeon E5-2400 series processors Up to two 2.4GHz (6-core) or 2.3GHz (8- core) processors Memory: 12 DIMMs (up to 192GB), up to 1600Mhz Hard disks: 2 x 3.5" HS HDDs 2 x 1Gb/sec Ethernet ports, 10GbE network port LSI 9207-8e SAS HBA Networking 10GbE nonblocking network switch GbE network switch OS Red Hat Enterprise Linux Server 6.2 (x86_64) or later Hadoop typically requires a Linux distribution Hadoop distribution Cloudera Distribution for Hadoop Cloudera Enterprise Core 4.1.3 Cloudera Manager 4.1.3 13 NetApp Solutions for Hadoop Reference Architecture

Component Product or Solution Details Hortonworks Data Platform 1.3 Management software NetApp SANtricity 10.84 SANtricity software is bundled with E- Series storage arrays 6 Other Supported Features, Products, and Protocols Although NetApp recommends the validated configuration in Table 1, individual customer requirements might call for different components. The components that are supported by NetApp and its partners are listed in Table 2. Table 2) Components supported by NetApp and its partners. Component or Feature Other Supported Options Details Storage arrays E5412 E5424 E27xx E55xx EF5xx Other E5400 storage arrays are also supported, as are the E2700 and E5500 series. Disk drives and type 12, 24, 60 1TB, 2TB 12 and 24 drives are supported as well as 1TB, 2TB, and 4TB drives. Protocols Other Hadoop distributions Fibre Channel ISCSI InfiniBand MapR BigInsights Intel Any Apache-compatible Hadoop distribution is supported. 7 Conclusion With big data growing from its roots in Internet companies to becoming more established in data centers, businesses are turning to Hadoop to help them ingest, store, manage, and analyze all the various data they are collecting. Typically, organizations that are deploying or considering Hadoop use traditional server-based storage in an internal DAS configuration. However, this can come with challenges in data centers because internal DAS might be not as reliable for an organization s SLAs, harder to scale, and not as flexible. The NetApp Solutions for Hadoop reference design employs the external or managed DAS storage, which provides dedicated storage, with higher scalability and reliability. There is also architectural flexibility by decoupling compute nodes from storage to scale one without needing to scale the other, being able to choose any NetApp E-Series and being able to use most servers and networking technologies. There is also freedom to choose The NetApp solutions for Hadoop reference architectures are validated to help customers get started with Hadoop or deploy Hadoop in production so they can begin mining their data for insights. 14 NetApp Solutions for Hadoop Reference Architecture

Additional Information To learn about NetApp Hadoop solutions, visit http://www.netapp.com/hadoop. Additional information about the base reference design is discussed on this website, and it also provides information about FlexPod Select with Hadoop, which is the validated base reference architecture with Cisco servers and networking. There are two Cisco Validated Designs for FlexPod Select with Hadoop: FlexPod Select with Cloudera Enterprise FlexPod Select with Hortonworks Data Platform If you want to contact NetApp sales or a partner to either investigate NetApp solutions further or start a pilot or proof of concept, click here. More information on the Hadoop project, technologies, and ecosystem can be found at www.hadoop.apache.org. You can also contact the authors at iyerv@netapp.com or gustav.horn@netapp.com. Version History Version Date Document Version History Version 1.0 April 2014 Initial release Refer to the Interoperability Matrix Tool (IMT) on the NetApp Support site to validate that the exact product and feature versions described in this document are supported for your specific environment. The NetApp IMT defines the product components and versions that can be used to construct configurations that are supported by NetApp. Specific results depend on each customer's installation in accordance with published specifications. NetApp provides no representations or warranties regarding the accuracy, reliability, or serviceability of any information or recommendations provided in this publication, or with respect to any results that may be obtained by the use of the information or observance of any recommendations provided herein. The information in this document is distributed AS IS, and the use of this information or the implementation of any recommendations or techniques herein is a customer s responsibility and depends on the customer s ability to evaluate and integrate them into the customer s operational environment. This document and the information contained herein may be used solely in connection with the NetApp products discussed in this document. 2014 NetApp, Inc. All rights reserved. No portions of this document may be reproduced without prior written consent of NetApp, Inc. Specifications are subject to change without notice. NetApp, the NetApp logo, Go further, faster, FlexPod, and SANtricity are trademarks or registered trademarks of NetApp, Inc. in the United States and/or other countries. Intel and Xeon are registered 15 NetApp Solutions for trademarks Hadoop of Reference Intel Corporation. Architecture Linux is a registered trademark of Linus Torvalds. Cisco is a registered trademark of Cisco Systems, Inc. All other brands or products are trademarks or registered trademarks of their respective holders and should be treated as such. WP-7196-0414