TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

Similar documents
TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

World s most advanced data center accelerator for PCIe-based servers

TESLA V100 PERFORMANCE GUIDE. Life Sciences Applications

NVIDIA DGX SYSTEMS PURPOSE-BUILT FOR AI

TESLA P100 PERFORMANCE GUIDE. HPC and Deep Learning Applications

ACCELERATED COMPUTING: THE PATH FORWARD. Jen-Hsun Huang, Co-Founder and CEO, NVIDIA SC15 Nov. 16, 2015

GPU ACCELERATED COMPUTING. 1 st AlsaCalcul GPU Challenge, 14-Jun-2016, Strasbourg Frédéric Parienté, Tesla Accelerated Computing, NVIDIA Corporation

unleashed the future Intel Xeon Scalable Processors for High Performance Computing Alexey Belogortsev Field Application Engineer

IBM POWER SYSTEMS: YOUR UNFAIR ADVANTAGE

Maximize automotive simulation productivity with ANSYS HPC and NVIDIA GPUs

TESLA P100 PERFORMANCE GUIDE. Deep Learning and HPC Applications

DataON and Intel Select Hyper-Converged Infrastructure (HCI) Maximizes IOPS Performance for Windows Server Software-Defined Storage

IBM Power Advanced Compute (AC) AC922 Server

IBM Power AC922 Server

How Architecture Design Can Lower Hyperconverged Infrastructure (HCI) Total Cost of Ownership (TCO)

Making Supercomputing More Available and Accessible Windows HPC Server 2008 R2 Beta 2 Microsoft High Performance Computing April, 2010

IBM Power Systems: Open innovation to put data to work Dexter Henderson Vice President IBM Power Systems

QLogic TrueScale InfiniBand and Teraflop Simulations

Micron and Hortonworks Power Advanced Big Data Solutions

HIGH-PERFORMANCE STORAGE FOR DISCOVERY THAT SOARS

Performance and Energy Efficiency of the 14 th Generation Dell PowerEdge Servers

Enhancing Analysis-Based Design with Quad-Core Intel Xeon Processor-Based Workstations

New Approach to Unstructured Data

Broadcast-Quality, High-Density HEVC Encoding with AMD EPYC Processors

HETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA

System Design of Kepler Based HPC Solutions. Saeed Iqbal, Shawn Gao and Kevin Tubbs HPC Global Solutions Engineering.

FUJITSU Server PRIMERGY CX400 M4 Workload-specific power in a modular form factor. 0 Copyright 2018 FUJITSU LIMITED

Evaluation Report: Improving SQL Server Database Performance with Dot Hill AssuredSAN 4824 Flash Upgrades

Composable Infrastructure for Public Cloud Service Providers

Shared Infrastructure: Scale-Out Advantages and Effects on TCO

Looking ahead with IBM i. 10+ year roadmap

Apache Spark Graph Performance with Memory1. February Page 1 of 13

Reduce costs and enhance user access with Lenovo Client Virtualization solutions

NVIDIA GRID. Ralph Stocker, GRID Sales Specialist, Central Europe

Timothy Lanfear, NVIDIA HPC

AMD Disaggregates the Server, Defines New Hyperscale Building Block

TESLA V100 PERFORMANCE GUIDE May 2018

WHITE PAPER. How VMware Virtualization Right-sizes IT Infrastructure to Reduce Power Consumption

HCI: Hyper-Converged Infrastructure

Why Converged Infrastructure?

VEXATA FOR ORACLE. Digital Business Demands Performance and Scale. Solution Brief

BUILD BETTER MICROSOFT SQL SERVER SOLUTIONS Sales Conversation Card

Why Converged Infrastructure?

Refining and redefining HPC storage

Leading in the compute era. Delivering the right compute, for the right workload, at the right economics every time.

Differentiate Your Business with Oracle PartnerNetwork. Specialized. Recognized by Oracle. Preferred by Customers.

Top 5 Reasons to Consider

W H I T E P A P E R U n l o c k i n g t h e P o w e r o f F l a s h w i t h t h e M C x - E n a b l e d N e x t - G e n e r a t i o n V N X

HP s Performance Oriented Datacenter

Analyzing the Economic Benefits of the HPE SimpliVity 380 All-flash Hyperconverged Portfolio

The Case for Virtualizing Your Oracle Database Deployment

Deploying Data Center Switching Solutions

Virtuozzo Hyperconverged Platform Uses Intel Optane SSDs to Accelerate Performance for Containers and VMs

White Paper. Platform9 ROI for Hybrid Clouds

RED HAT ENTERPRISE LINUX. STANDARDIZE & SAVE.

Data safety for digital business. Veritas Backup Exec WHITE PAPER. One solution for hybrid, physical, and virtual environments.

IBM Storwize V7000 TCO White Paper:

Lenovo Validated Designs

Optimizing Apache Spark with Memory1. July Page 1 of 14

Lenovo Data Center Group. Define a different future

Cloud Computing: Making the Right Choice for Your Organization

34% DOING MORE WITH LESS How Red Hat Enterprise Linux shrinks total cost of ownership (TCO) compared to Windows. I n a study measuring

Cray XC Scalability and the Aries Network Tony Ford

Performance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA

Dell EMC Hyperconverged Portfolio: Solutions that Cover the Use Case Spectrum

Shared infrastructure: scale-out advantages and effects on TCO

HPE ProLiant ML350 Gen P 16GB-R E208i-a 8SFF 1x800W RPS Solution Server (P04674-S01)

APPLYING DATA CENTER INFRASTRUCTURE MANAGEMENT IN COLOCATION DATA CENTERS

IBM Power Systems HPC Cluster

IT Consulting and Implementation Services

Expand In-Memory Capacity at a Fraction of the Cost of DRAM: AMD EPYCTM and Ultrastar

SUPERMICRO, VEXATA AND INTEL ENABLING NEW LEVELS PERFORMANCE AND EFFICIENCY FOR REAL-TIME DATA ANALYTICS FOR SQL DATA WAREHOUSE DEPLOYMENTS

DIGITALGLOBE ENHANCES PRODUCTIVITY

Virtualization strategy for mid-sized businesses. IBM, Intel and VMware virtualization benefits for mid-sized businesses

Trends in Next Generation Data Center Infrastructure. Summary Results January 2018

AN INTRODUCTION TO CLUSTER COMPUTING

Benefits of SD-WAN to the Distributed Enterprise

Organizational Update: December 2015

CAPEX Savings and Performance Advantages of BigTwin s Memory Optimization

Cisco Unified Computing System Delivering on Cisco's Unified Computing Vision

HPE ProLiant ML110 Gen10 Server

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

Global Headquarters: 5 Speen Street Framingham, MA USA P F

Is remote GPU virtualization useful? Federico Silla Technical University of Valencia Spain

Converged Infrastructure Matures And Proves Its Value

Remote GPU virtualization: pros and cons of a recent technology. Federico Silla Technical University of Valencia Spain

Driving Data Warehousing with iomemory

ACCELERATE YOUR ANALYTICS GAME WITH ORACLE SOLUTIONS ON PURE STORAGE

AMD EPYC Empowers Single-Socket Servers

A Breakthrough in Non-Volatile Memory Technology FUJITSU LIMITED

DATA CENTRE SOLUTIONS

Ten things hyperconvergence can do for you

Cisco CloudCenter Use Case Summary

Hybrid Cloud 1. ebookiness created by the HPE Europe Division of Ingram Micro

High performance and functionality

Microsoft Exchange Server 2010 workload optimization on the new IBM PureFlex System

Find the right platform for your server needs

BUSTED! 5 COMMON MYTHS OF MODERN INFRASTRUCTURE. These Common Misconceptions Could Be Holding You Back

Renovating your storage infrastructure for Cloud era

Transcription:

TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

Table of Contents: The Accelerated Data Center Optimizing Data Center Productivity Same Throughput with Fewer Server Nodes Aren t GPU-Accelerated Servers More Expensive? What About Operational Cost? What if the CPU is Given Away at No Cost? Maximizing Budget and Throughput Lower Cost with Acceleration Democratizing the Supercomputer 3 6 7 8 10 11 12 13 14 ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 2

Accelerated computing is revolutionizing the economics of the data center. HPC, enterprise, and hyperscale customers deploy accelerated servers because GPUs deliver unprecedented cost savings for their data center. This whitepaper provides an analysis on how accelerators like the NVIDIA Tesla V100 can lower data center cost by up to 45%. The Accelerated Data Center Today, data centers are built by interconnecting many COTS (commercial off-the-shelf) technologies. Customers typically deploy the most costeffective system by making trade-offs between commodity components like CPU, memory, and networking. However, cost savings are incremental. Accelerators fundamentally change the economics of the data center, producing application performance gains that are no longer incremental. A single GPU-accelerated server can replace over 100 CPU-only servers. The number of CPU-only servers replaced by a GPU-accelerated sever is called the node replacement factor (NRF). The NRF will vary based on the applications running on the server. One GPU-accelerated servers (with 4xV100) running a balanced workload allocation of these popular HPC applications replaces 31 CPU-only servers (NRF= 31X). However, the workload mix of supercomputing sites include more AI workloads which typically have an NRF of over 40X. (See the HPC Application Performance Guide for details) Figure-1: A mixed workload of popular HPC applications has an NRF of 31X. Nvidia Tesla Accelerator Performance By Application 120X 109X 100X 80X 74X 79X Speedup 60X 40X 20X 26X 12X 46X Mixed HPC Avg Node Replace Rate: 31X 23X 0X AMBER Chroma GTC LAMMPS MILC NAMD SPECFEM3D CPU Server: Dual Xeon Gold 6140 @ 2.30GHz, GPU Servers: same CPU server with 4X V100 PCIe Figure-1 ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 3

A recent survey by Intersect360 Research, a leading HPC analyst firm, showed that 15 of the top 15, and 20 of the top 25 most popular HPC applications support GPU acceleration today. Figure-2: Many of the most popular applications in HPC and deep learning are GPU-accelerated, to deliver unprecedented productivity and cost savings in the data center. 2017 Intersect360 Survey of Top Apps TOP 15 HPC APPS 100% Accelerated TOP 50 HPC APPS 70% Accelerated Intersect 360, Nov 2017, HPC Application Support for GPU Computing Figure-2 Now, with over 580 GPU-accelerated applications, and more applications being added every day, the question is no longer if GPUs should be deployed in the data center, but how many. This broad coverage of GPU-accelerated applications will help boost data center throughput and utilization, resulting in dramatic improvements in cost savings. Figure-3: With over 580 GPU-accelerated applications, IT managers can optimize data center throughput and optimization. Over 580 HPC Applications are GPU-Accelerated 700 Number of GPU-Accelerated Apps 600 500 400 300 200 100 259 319 400 470 581 0 2014 2015 2016 2017 2018 Figure-3 ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 4

In addition to increasing the number of GPU-accelerated applications, NVIDIA continually optimizes each component of the software/hardware stack to improve end-to-end application performance over time. In Figure 4 we show the average NRF for a mixed-hpc application workload (shown in Figure 1) has increased from 6X to 31X in just 3 years. Figure-4: Tesla platform performance improvement over time. Node Reduction Factor (NRF) increases from 6X to 31X in just 3 years NRF Of Mixed HPC Workload From 2016 To 2018 40X 35X 30X 31X 25X Speedup 20X 15X 12X 10X 5X 6X 0X 2016 K80 Haswell 2017 P100 Broadwell 2018 V100 Skylake CPU Server: Dual-CPU, GPU Servers: 4x PCIe GPUs Figure-4 ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 5

A manufacturing facility is evaluated based on its production or throughput, and the data center is no exception. Optimizing Data Center Productivity The productivity of an infrastructure is often measured by its throughput. In manufacturing, the primary metric driving profitability is the number of goods produced per day. In cloud services, users pay based on tiers of data throughput, measured in megabytes per second. The data center is no different. Throughput Is A Key Measure Of Data Center Productivity MANUFACTURING Input Raw Goods Input Compute Jobs Figure-5 Data center throughput is measured by the amount of work completed in a given period of time (i.e. number of jobs per day or per month). While this architecture is complex, users typically abstract away all the system complexities into a simple working model; they submit job requests into a black box through a job scheduler and expect results soon after. In high-performance computing, researchers rely on velocity of output of the data center for discoveries and insights. Higher data center throughput means more scientific discoveries are delivered to researchers every day. In a web services company, thousands of consumers using various types of devices may request live video streaming of a trending event. Higher throughput means a better user experience. Throughput is a key measure of data center productivity. ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 6

The CPU-Only data center has 1,000 CPU-only nodes processing 1,000 jobs per day. The accelerated data center has 300 CPU-only nodes and 35 nodes with Tesla V100 accelerators, delivering the same throughput as the CPU-Only data center with 67% less nodes. Same Throughput with Fewer Server Nodes To illustrate how throughput and cost savings are related, let s assume there are two data centers. The CPU-Only data center is comprised of traditional CPU servers and the accelerated data center is comprised of a mix of traditional CPU servers and GPU-accelerated servers. Each node is dual-socket CPU design, while GPU-accelerated nodes have two NVIDIA Tesla V100 accelerators attached to the node. In terms of workload profile for both data centers, we re assuming 70% of the jobs are based on applications that support GPU computing. In this whitepaper, we ll assume that a single CPU node can process one unit of work, or job, per day. So the CPU-Only data center has 1,000 nodes capable of processing 1,000 jobs. Let s take a look at the accelerated data center. Because 70% of jobs support GPU computing, 700 jobs in the queue can run on GPU-accelerated nodes while 300 jobs should run on CPU-only nodes. With a conservative assumption that GPU-enabled jobs run 20X faster on a Tesla V100 node compared to a CPU-only node, only 35 accelerated nodes are needed to deliver 700 jobs per day. 300 CPU nodes are required for the remaining jobs in the queue, for a total of 335 server nodes. The accelerated data center delivers the same productivity with 67% less servers, racks, and networking equipment. This translates into tremendous savings in both acquisition cost as well as operation cost due to lower power and smaller physical space requirements. ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 7

Aren t GPU-Accelerated Servers More Expensive? Accelerators add cost to a node, so customers often make the mistake of concluding that the GPU-accelerated solution is more expensive. To analyze the cost impact of adding accelerators, let s start with a comparison of server node cost. SERVER COSTS CPU-ONLY NODE (Dual Socket CPU) CPU (x2) $2,000 $2,000 GPU (x4) $8,500 NIC, Memory, Misc. Cost $4,000 $4,000 Core Networking Per Node $1,000 $1,000 Total Node Cost $9,000 $44,000 Table-1: Breakdown of CPU-only and GPU-accelerated node cost. ACCELERATED NODE (4x Tesla V100) A single CPU socket costs $2,000 with other necessary components, like NICs and DDR4 memory, and core networking components that cost $5,000, is a total of $9,000 per CPU-only node. By adding four Tesla V100 accelerators to the same node design, the node cost now totals $44,000. While node cost is higher with GPUs, nodes cannot operate without other data center technologies like storage, software, and services. The typical breakdown of data center cost is as follows: DATACENTER TECHNOLOGIES % OF SYSTEM ACQUISITION BUDGET Servers 70% Storage 20% Software and Services 10% Total Acquisition Cost 100% Table-2: Typical system acquisition budget allocation for the data center. ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 8

Using this breakdown, the CPU-Only data center would need to budget $9M for server nodes and approximately $4.5M for storage, software, and services. The accelerated data center requires a smaller budget for server nodes and networking, since there are fewer nodes to interconnect. The budget for software and services also decreases due to fewer nodes and smaller overall system costs. However, we ll assume that the budget for storage is the same as the CPU-Only data center. ACQUISITION COSTS THE CPU-ONLY DATA THE ACCELERATED CPU Nodes $9,000 x 1000 Nodes $9,000 x 300 Nodes Tesla V100 Nodes $44,000 x 35 Nodes Servers $9.0M $4.2M Storage $3.0M $3.0M Software and Services $1.5M $.9M Total Acquisition Cost $13.5M $8.1M Table-3: The accelerated data center with Tesla V100 reduces system acquisition cost by 40% compared to the CPU-Only data center. While the customer needs $13.5M to deploy the CPU-Only data center, they only need $8.1M when some of the nodes are accelerated as in the accelerated data center. The end result is acquisition cost savings of $5.4M (40%) when choosing the accelerated data center. Figure-6: 40% cost savings with the accelerated data center. System Cost For Same Throughput The CPU-Only Data Center The Accelerated Data Center $8.1M $13.5M $0 $2 $4 $6 $8 $10 $12 $14 System Acqusition Cost (Millions) Figure-6 ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 9

What About Operational Cost? Once the servers are deployed in the data center they will also generate ongoing operational costs such as power consumption and cooling infrastructure. Today, many data center managers will assign a value that ranges from $100-$300/kw/Month to capture these operational costs. For this example we ll assume $200/kW/Month for the 3-year life of each server. OPERATIONAL COSTS THE CPU-ONLY THE ACCELERATED CPU Nodes 1000 Nodes 300 Nodes Tesla V100 Nodes 35 Nodes Power Consumption per Node 600W 1,700W 3-year Operational Cost ($200/kW*36) $4.3M $1.7M Table-4: 3-year operational costs, the accelerated data center with Tesla V100 is still saves 60%. TOTAL COSTS THE CPU-ONLY Acquisition Cost $13.5M $8.1M 3-year Operational Cost ($200/kW*36) $4.3M $1.7M Total Cost $17.8M $9.8M THE ACCELERATED Table-5: Total data center cost over 3 years, the accelerated data center with Tesla V100 saves 45%. Based on this analysis, there is an additional $2.6M (57%) savings in 3-year operational costs, resulting in a total 3-year cost savings of $8.0M (45%). ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 10

What if the CPU is Given Away at No Cost? In some situations, the CPU may be discounted in an attempt to stay competitive with a GPU-accelerated solution. So let s take an extreme, unlikely case where the CPU is free for the CPU-Only data center, but full cost for the accelerated data center. ACQUISITION COSTS THE CPU-ONLY DATA CENTER (Free CPU) THE ACCELERATED CPU Nodes $5,000 x 1000 Nodes $9,000 x 300 Nodes Tesla V100 Nodes $44,000 x 35 Nodes Servers $5.0M $4.2M Storage $3.0M $3.0M Software and Services $1.5M $0.9M Total Acquisition Cost $9.5M $8.1M TOTAL DATA CENTRER COSTS THE CPU-ONLY Acquisition Cost $9.5M $8.1M 3-year Operational Cost ($200/kW*36) $4.3M $1.7M Total Cost $13.8M $9.8M THE ACCELERATED Table-6: Even if CPUs were sold at zero-cost in CPU-Only data center, the accelerated data center with Tesla V100 is still saves 29%. Node costs are reduced by 44% to $5,000 and the overall acquisition cost is decreased by 30% to $9.5 million. However, the CPU-Only data center still requires $1.4M (17%) more than the accelerated data center. Additionally, when you include the operational costs discussed above, the accelerated data center is $4.0M lower in overall cost (29%). ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 11

Maximizing Budget and Throughput If a customer has a fixed budget that must be spent, Tesla V100 offers unprecedented ROI by maximizing throughput. With 70% of the top applications already leveraging GPU acceleration, and more applications on the way, many customers choose to deploy more GPUs into the data center. With $5.4M of acquisition cost savings generated by the accelerated data center, the IT manager can also decide to purchase more GPU nodes to introduce new GPU-accelerated applications such as Machine Learning and Deep Learning to the data center workload mix. Let s call this new data center, the max-accelerated data center, which contains a mix of CPU-only nodes and additional GPU-accelerated nodes to run new AI workloads running over 40X faster on GPU-accelerated servers. COST THE CPU-ONLY THE ACCELERATED CPU Nodes $9,000 x 1000 Nodes $9,000 x 300 Nodes Tesla V100 Nodes $44,000 x 143 Nodes Servers $9.0M $9.0M Storage $3.0M $3.0M Software and Services $1.5M $1.5M Total Acquisition Cost $13.5M $13.5M Total Data Center Throughput 1000 Jobs/Day 5327 Jobs/Day Table-7: The max-accelerated data center with Tesla V100 delivers over 5x higher throughput compared to the CPU-Only data center. OPERATIONAL COSTS THE CPU-ONLY THE ACCELERATED CPU Nodes 1000 Nodes 300 Nodes Tesla V100 Nodes 143 Nodes Power Consumption per Node 600W 1,700W 3-year Operational Cost ($200/kW*36) $4.3M $3.0M Table-8: 3-year operational costs, the Max-accelerated data center with Tesla V100 still saves $1.3M. IT managers can deploy 108 more GPU nodes with the cost savings made possible by Tesla V100 GPUs. The max-accelerated data center also enjoys lower operational costs due to lower number of nodes. With 108 new GPU nodes producing a 40X throughput increase for AI applications, the max-accelerated data center delivers roughly 5300 jobs per day, an over 5X increase in throughput compared to the CPU-Only data center. That adds up to a 3-year operational cost savings of $1.3M. ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 12

Lower Cost with Acceleration IT managers care about cost savings. The budget is never big enough to cover all the programs and equipment required to keep the organization working smoothly, so any cost savings is a welcome relief. With Tesla V100, IT managers have slashed their data center costs. Figure-7: GPU-Accelerated data centers reduce costs significantly and deliver the best throughput per dollar. (Data in chart normalized to the CPU-Only data center) Lower Cost with Accelerations 6X 5X 4X 5.8X Max-Accelerated Datacenter Throughput 3X 2X 1X 1.0X 1.8X 0X CPU Only Datacenter Accelerated Datacenter 0X 1X 2X 3X 4X 5X 6X Throughput Per Dollar Figure-7 In this whitepaper, we used three examples of data centers that customers deploy today. Compared to the CPU-Only data center, Tesla V100 in the accelerated data center results in 40% reduction in acquisition costs, saving $4.8 million in an overall budget of $13.5 million. For customers who want to maximize productivity, Tesla V100 in the max-accelerated data center delivers over 5X increase in overall productivity for the same $13.5 million acquisition budget. ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING TECHNICAL OVERVIEW 13

Democratizing the Supercomputer Researchers and engineers in HPC, Hyper scale, and Fortune 100 companies work with massive amounts of data as data center costs are growing. GPU-accelerated computing gives them better performance/ throughput while also helping manage their costs. Accelerated computing democratizes supercomputing, making it affordable for more researchers, scientists, and enterprise companies to deploy the system they need. Now, a university team focused on finding a cure for cancer or a Fortune 100 company committed to innovation can each afford computing power previously reserved for supercomputing facilities. 2018 NVIDIA Corporation. All rights reserved. NVIDIA, the NVIDIA logo, and Tesla are trademarks and/or registered trademarks of NVIDIA Corporation. All company and product names are trademarks or registered trademarks of the respective owners with which they are associated.