NVMe SSDs Future-proof Apache Cassandra

Similar documents
NVMe SSDs A New Benchmark for OLTP Performance

Micron and Hortonworks Power Advanced Big Data Solutions

Micron Quad-Level Cell Technology Brings Affordable Solid State Storage to More Applications

Accelerate Database Performance and Reduce Response Times in MongoDB Humongous Environments with the LSI Nytro MegaRAID Flash Accelerator Card

Three Paths to Better Business Decisions

Aerospike Scales with Google Cloud Platform

Accelerating Enterprise Search with Fusion iomemory PCIe Application Accelerators

Accelerating Big Data: Using SanDisk SSDs for Apache HBase Workloads

Dell PowerEdge R730xd Servers with Samsung SM1715 NVMe Drives Powers the Aerospike Fraud Prevention Benchmark

Technical Note. Improving Random Read Performance Using Micron's SNAP READ Operation. Introduction. TN-2993: SNAP READ Operation.

A Comparative Study of Microsoft Exchange 2010 on Dell PowerEdge R720xd with Exchange 2007 on Dell PowerEdge R510

Fusion iomemory PCIe Solutions from SanDisk and Sqrll make Accumulo Hypersonic

Apache Spark Graph Performance with Memory1. February Page 1 of 13

Virtualizing SQL Server 2008 Using EMC VNX Series and VMware vsphere 4.1. Reference Architecture

Evaluation Report: Improving SQL Server Database Performance with Dot Hill AssuredSAN 4824 Flash Upgrades

EMC Backup and Recovery for Microsoft SQL Server

Intel Solid State Drive Data Center Family for PCIe* in Baidu s Data Center Environment

NoSQL Performance Test

Upgrade to Microsoft SQL Server 2016 with Dell EMC Infrastructure

Dell PowerEdge R720xd with PERC H710P: A Balanced Configuration for Microsoft Exchange 2010 Solutions

A High-Performance Storage and Ultra- High-Speed File Transfer Solution for Collaborative Life Sciences Research

Cloud Computing with FPGA-based NVMe SSDs

Client vs. Enterprise SSDs

The Impact of SSD Selection on SQL Server Performance. Solution Brief. Understanding the differences in NVMe and SATA SSD throughput

IBM System Storage DCS3700

Ceph in a Flash. Micron s Adventures in All-Flash Ceph Storage. Ryan Meredith & Brad Spiers, Micron Principal Solutions Engineer and Architect

CIS 601 Graduate Seminar. Dr. Sunnie S. Chung Dhruv Patel ( ) Kalpesh Sharma ( )

It s Time to Move Your Critical Data to SSDs Introduction

EMC Virtual Infrastructure for Microsoft Exchange 2010 Enabled by EMC Symmetrix VMAX, VMware vsphere 4, and Replication Manager

Lenovo Database Configuration

Lenovo Database Configuration for Microsoft SQL Server TB

Dell Reference Configuration for Large Oracle Database Deployments on Dell EqualLogic Storage

Huge Benefits With All-Flash and vsan 6.5

Architecture of a Real-Time Operational DBMS

EMC Backup and Recovery for Microsoft Exchange 2007 SP1. Enabled by EMC CLARiiON CX4-120, Replication Manager, and VMware ESX Server 3.

THE SUMMARY. CLUSTER SERIES - pg. 3. ULTRA SERIES - pg. 5. EXTREME SERIES - pg. 9

Using the ThinkServer SA120 Disk Array

Dell EMC SAP HANA Appliance Backup and Restore Performance with Dell EMC Data Domain

EMC Backup and Recovery for Microsoft Exchange 2007

Comparing UFS and NVMe Storage Stack and System-Level Performance in Embedded Systems

Implementing SQL Server 2016 with Microsoft Storage Spaces Direct on Dell EMC PowerEdge R730xd

An Oracle Technical White Paper October Sizing Guide for Single Click Configurations of Oracle s MySQL on Sun Fire x86 Servers

High-Performance Lustre with Maximum Data Assurance

Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads

Evaluation Report: HP StoreFabric SN1000E 16Gb Fibre Channel HBA

Seagate Enterprise SATA SSD with DuraWrite Technology Competitive Evaluation

Deploying Apache Cassandra on Oracle Cloud Infrastructure Quick Start White Paper October 2016 Version 1.0

Best Practices for SSD Performance Measurement

Fast and Easy Persistent Storage for Docker* Containers with Storidge and Intel

NVMFS: A New File System Designed Specifically to Take Advantage of Nonvolatile Memory

Dell EMC SCv3020 7,000 Mailbox Exchange 2016 Resiliency Storage Solution using 7.2K drives

Microsoft Office SharePoint Server 2007

NOSQL DATABASE CLOUD SERVICE. Flexible Data Models. Zero Administration. Automatic Scaling.

Performance of popular open source databases for HEP related computing problems

HP SmartCache technology

Dell DL1000 Appliance Interoperability Guide

SSDs Going Mainstream Do you believe? Tom Rampone Intel Vice President/General Manager Intel NAND Solutions Group

TECHNICAL OVERVIEW OF NEW AND IMPROVED FEATURES OF EMC ISILON ONEFS 7.1.1

White paper ETERNUS Extreme Cache Performance and Use

Top 5 Reasons to Consider

Lenovo Database Configuration

MongoDB on Kaminario K2

REFERENCE ARCHITECTURE Microsoft SQL Server 2016 Data Warehouse Fast Track. FlashStack 70TB Solution with Cisco UCS and Pure Storage FlashArray//X

SUPERMICRO, VEXATA AND INTEL ENABLING NEW LEVELS PERFORMANCE AND EFFICIENCY FOR REAL-TIME DATA ANALYTICS FOR SQL DATA WAREHOUSE DEPLOYMENTS

EMC Business Continuity for Microsoft Applications

Over-Provisioning the Micron 1100 SSD for Data Center Applications

Nimble Storage Adaptive Flash

EMC ISILON S-SERIES. Specifications. EMC Isilon S200. EMC Isilon S210 ARCHITECTURE

HP Z Turbo Drive G2 PCIe SSD

Drecom Accelerates Anytime, Anywhere Gaming

Supermicro All-Flash NVMe Solution for Ceph Storage Cluster

Micron NVMe in the Open19 Platform: Low-Latency Storage at the Edge.

All-NVMe Performance Deep Dive Into Ceph + Sneak Preview of QLC + NVMe Ceph

Dell EMC SC Series SC5020 9,000 Mailbox Exchange 2016 Resiliency Storage Solution using 7.2K Drives

Deep Learning Performance and Cost Evaluation

EMC XTREMCACHE ACCELERATES VIRTUALIZED ORACLE

Create a Flexible, Scalable High-Performance Storage Cluster with WekaIO Matrix

Veeam Availability Solution for Cisco UCS: Designed for Virtualized Environments. Solution Overview Cisco Public

Dell EMC All-Flash solutions are powered by Intel Xeon processors. Learn more at DellEMC.com/All-Flash

Agenda. AWS Database Services Traditional vs AWS Data services model Amazon RDS Redshift DynamoDB ElastiCache

HPE SimpliVity. The new powerhouse in hyperconvergence. Boštjan Dolinar HPE. Maribor Lancom

Data Protection for Cisco HyperFlex with Veeam Availability Suite. Solution Overview Cisco Public

NAMD Performance Benchmark and Profiling. January 2015

Functional Testing of SQL Server on Kaminario K2 Storage

Performance Benefits of Running RocksDB on Samsung NVMe SSDs

EMC Virtual Infrastructure for Microsoft Applications Data Center Solution

THESUMMARY. ARKSERIES - pg. 3. ULTRASERIES - pg. 5. EXTREMESERIES - pg. 9

Altair OptiStruct 13.0 Performance Benchmark and Profiling. May 2015

Scaling Internet TV Content Delivery ALEX GUTARIN DIRECTOR OF ENGINEERING, NETFLIX

Lenovo Database Configuration Guide

Selecting Server Hardware for FlashGrid Deployment

Dell PowerVault NX Windows NAS Series Configuration Guide

Data-Centric Innovation Summit ALPER ILKBAHAR VICE PRESIDENT & GENERAL MANAGER MEMORY & STORAGE SOLUTIONS, DATA CENTER GROUP

EMC VFCache. Performance. Intelligence. Protection. #VFCache. Copyright 2012 EMC Corporation. All rights reserved.

Deep Learning Performance and Cost Evaluation

ADVANCED IN-MEMORY COMPUTING USING SUPERMICRO MEMX SOLUTION

Flash Storage Complementing a Data Lake for Real-Time Insight

Oracle Database Exadata Cloud Service Exadata Performance, Cloud Simplicity DATABASE CLOUD SERVICE

New Approach to Unstructured Data

Hyper-converged storage for Oracle RAC based on NVMe SSDs and standard x86 servers

Transcription:

NVMe SSDs Future-proof Apache Cassandra Get More Insight from Datasets Too Large to Fit into Memory Overview When we scale a database either locally or in the cloud performance 1 is imperative. Without massive performance, a massive-scale database is little more than an active archive. When an entire data set is small and fits into memory (DRAM), performance is straight-forward and storage system capability is less important. However, with immense data growth, a dwindling percentage of data affordably fits into memory. By building with SSDs, we can future-proof Apache Cassandra deployments to perform soundly as active data sets grow, extending well beyond memory capacity. Combined with the constant demand for faster and more detailed analytics, we have arrived at a data-driven crossroads: We need high performance, high capacity and affordability. Fast Facts - A four-node cluster using a single NVMe SSD eclipsed the capability of a multidrive legacy (HDD) fournode cluster across multiple workloads and thread counts - A single NVMe SSD per node configuration measured up to 31X better performance, with more consistent, lower latency Cassandra combined with NVMe SSDs can help. Cassandra s ability to support massive scaling, combined with multiterabyte, high IOPS NVMe SSDs, builds highcapacity NoSQL platforms offering extreme capacity, extreme agility and extreme capability. This technical brief highlights the performance advantages we measured when we compared two 4-node Cassandra clusters: one built using legacy hard disk drives (HDDs); the second build using NVMe SSDs. We also explore some implications of these results. Due to the broad range of Cassandra deployments, we tested multiple workloads and multiple thread counts. You may find some results more relevant than others for your deployment. 1

NVMe SSDs Meet Growing Demands When we built Cassandra nodes with legacy HDD storage, we scaled out by adding more nodes to the cluster. We scaled up by upgrading to larger drives. Sometimes we did both. Adding more legacy nodes was effective (to a point), but it quickly became unwieldy. We gained capacity and a bit more performance, but as we added to the clusters, they became larger and more complex, consuming more rack space and support resources. Upgrading to larger HDDs was somewhat effective (also to a point) since we got more capacity per node and more capacity per cluster, but these upgrades rarely augmented cluster performance. With both techniques, performance stagnated while demand grew. High capacity, lightning-quick NVMe SSDs are changing the design rules. With single SSD capacities measured in terabytes (TB), throughput in gigabytes per second (GB/s) and IOPS in hundreds of thousands 2, high-capacity NVMe SSDs enable new design opportunities and performance thresholds. We used the Yahoo! Cloud Serving Benchmark (YCSB) workloads A D and F 3 to compare two 4-node Cassandra test clusters: one built with NVMe SSDs and the other built with multiple legacy HDDs. Note: Due to the broad range of Cassandra deployments, we tested multiple thread counts from 48 to 480. See the How We Tested section for details. NVMe Clusters Build Capacity and Results As you plan your next high-capacity, high-demand Cassandra cluster, NVME SSDs can support amazing capacity and provide compelling results. Using a single NVMe SSD, each node in our SSD test cluster stores about 7.68TB. With six 15K RPM 300GB drives (RAID 0), our HDD test cluster stores about 1.8 TB per node. SSD Test Cluster: One 7.68GB SSD with NVMe per node (4 cluster nodes) Legacy Test Cluster: Six 300GB 15K RPM HDDs, RAID 0 poer node (4 cluster nodes) With the same number of nodes and a single SSD in each node, the NVMe SSD test cluster offers a 4X capacity increase. We also measured a tremendous increase in performance over all the workloads and thread counts tested, ranging from a low of about 2X to a high of 31X, along with lower and more consistent latency. Figure 1 shows YCSB performance for each configuration. 2

80,000 Operations per Second by Workload, Thread Count and Drive SSD with NVMe Legacy (HDD) 70,000 60,000 50,000 Operatons/second 40,000 30,000 20,000 10,000 0 48 96 192 240 480 48 96 192 240 480 48 96 192 240 480 48 96 192 240 480 48 96 192 240 480 Workload A Workload B Workload C Workload D Workload F Figure 1: Relative Performance NVMe Clusters Provide More Consistent Read Response Since many Cassandra deployments rely heavily on fast, consistent read responses, we compared the 99 th percentile read response times for each test cluster, workload and thread count. Figure 2 shows the results for each configuration. 1,400,000.00 99th Percentile Read Latency (µs) SSD with NVMe Legacy (HDD) 1,200,000.00 1,000,000.00 µs 800,000.00 600,000.00 400,000.00 200,000.00 0.00 48 96 192 240 480 48 96 192 240 480 48 96 192 240 480 48 96 192 240 480 48 96 192 240 480 Workload A Workload B Workload C Workload D Workload F Figure 2: Relative Read Responsiveness 3

The Bottom Line High-capacity, high-performance NVMe SSDs can produce amazing results with Cassandra. Whether you are scaling your local or cloud-based Cassandra deployment for higher performance or faster, more consistent read responses, NVMe SSDs are a great option. We tested two clusters for database performance and read responsiveness across multiple workloads and thread counts. We built a legacy cluster using six 300GB 15K RPM HDDs (RAID 0) in each node and another cluster using a single 7.68TB NVMe SSD in each node. The results were amazing. The single SSD per node test cluster showed a tremendous increase in performance over all the workloads and thread counts tested, ranging from a low of about 2X up to a high of 31X. We also found that the SSD-based cluster read responses were much faster with far greater consistency despite using only one NVMe SSD in each node. We expect great performance when our data set fits into memory, but immense data growth means that smaller and smaller portions of that data affordably fit into memory. We are at a crossroads. Our demands drive us toward higher performance, and data growth drives us toward affordable capacity. When we combine these, the answer is clear: NVMe SSDs deliver Cassandra performance and capacity that s more approachable. 1. We use the terms database operations per second (OPS) and performance interchangeably in this paper. 2. Capacity, GB/s and IOPS vary by SSD. This paper focuses on our 7.68TB U.2 9200. Other NVMe SSD models and/or capacities may give different results. 3. We did not test YCSB workload E because it is not universally supported. Note: We tested with Apache Cassandra Community Edition 3.11.1. Each node was equipped with 2x Intel Xeon E5-2690 v3 12 core processors and 256GB RAM. 4

How We Tested Table 1 shows the tested configurations, types of storage devices used, the number and capacity of each as well as the number of nodes in each Cassandra test cluster. Table 2 shows the hardware and software configuration parameters used. Configuration Drive Type Drives per Node Capacity per Node Nodes per Cluster SSD with NVMe 7.68TB ECO 1 7.68TB 4 Legacy 300GB, 15K RPM 6 1.8TB 4 Table 1: Tested Configuration Capacities Component Test Cluster Configuration/Description Controller Database Storage OS Settings for Cassandra NVMe SSD Legacy NVMe SSD Legacy NVMe SSD Legacy Not used for database storage Broadcom SAS 9300-8i (HBA) 1x Micron 9200 ECO 7.68TB 2.5 SSD with NVMe 6x 15K RPM 300GB HDD RAID 0 (mdadm) /etc/security/limits.d/cassandra.conf: cassandra - memlock unlimited cassandra - nofile 100000 cassandra - nproc 32768 cassandra - as unlimited /etc/sysctl.conf: vm.max_map_count = 131072 /etc/security/limits.d/20-nproc.conf: * - nproc 32768 Table 2: Configuration Parameters Our test methodology approximates real-world deployments and uses for a Cassandra database. Although the test configuration is relatively small (four nodes in each cluster), Cassandra s scaling technology means these results are also relevant to larger deployments. Four nodes host the database. The replication factor for the database was set to 3 (there are three copies of the data and the cluster can sustain the loss of two data nodes and continue to function). The database is initially created by utilizing YCSB workload A s load parameter, which generated a dataset of approximately 1.6TB, far exceeding available DRAM (ensuring we measure storage system IO). The database is then backed up to a separate location on the server for quick reload of data between test runs. For each configuration under test, the database was restored from this backup, starting every test from a consistent state. Table 3 shows the percentage of data owned by each of the four nodes. Node Capacity Tokens Percent Owned Node01 368.81GB 256 74.1% Node02 362.62GB 256 72.9% Node03 389.00GB 256 78.1% Node04 373.42GB 256 74.9% Table 3: Data Distribution across Nodes 5

Table 4 shows the testing parameters used in the tested workloads. Parameter Value Description Threads 48, 96, 144, 240, 480 Database load Field Count 10 1K record size (standard) Record Count 500 million Number of database records Operation Count 50 million Dataset size within database Table 4: Test Parameters Dim_stat was used to capture statistics on the server running Apache Cassandra. It captures IOStat, VMStat, mpstat, network load, processor load, and several other statistics. Dim_stat was configured to capture statistics on a 10-second interval. Table 5 shows the IO profiles for tested YCSB workloads (additional details are available at YCSB Core Workloads). Name Type IO Profile A Update heavy 50% Read, 50% Write B Read mostly 95% Read, 5% Write C Read only 100% Read, 0% Write D Read latest 95% Read, 5% Insert F Read-modify-write (R/M/W) 50% Read, 50% R/M/W Table 5: Workloads micron.com This technical brief is published by Micron and has not been authorized, sponsored, or otherwise approved by the Apache Software Foundation. Products are warranted only to meet Micron s production data sheet specifications. Products, programs and specifications are subject to change without notice. Dates are estimates only. 2017 Micron Technology, Inc. All rights reserved. All information herein is provided on an "AS IS" basis without warranties of any kind. Micron and the Micron logo are trademarks of Micron Technology, Inc. Apache and Cassandra are trademarks of the Apache Software Foundation. All other trademarks are the property of their respective owners. Rev. A 12/17, CCM004-676576390-10911 6