NVDIA DGX Data Center Reference Design

Size: px
Start display at page:

Download "NVDIA DGX Data Center Reference Design"

Transcription

1 White Paper NVDIA DGX Data Center Reference Design Easy Deployment of DGX Servers for Deep Learning NVIDIA Corporation.

2 Contents Abstract ii 1. AI Workflow and Sizing 1 2. NVIDIA AI Software 2 3. DGX POD 5 4. DGX POD Installation and Management 8 5. Summary NVIDIA Corporation. i

3 Abstract The NVIDIA DGX family of supercomputers for artificial intelligence (AI) and the NVIDIA GPU Cloud (NGC) Deep Learning (DL) containers provide the world s best solution for AI software development. However, many organizations are unsure how to size, install, and manage a DGX infrastructure for large-scale AI development efforts. The DGX POD reference architecture provides a blueprint for easy sizing and deployment of DGX servers for large-scale DL. A DGX POD is a single data center rack containing up to nine NVIDIA DGX-1 servers, storage, networking, and NVIDIA AI software to support workgroups of AI developers. Multiple DGX PODs can be combined into a cluster to support larger workgroups. The DGX POD reference architecture is based on the NVIDIA DGX SATURNV AI supercomputer which has 1000 DGX-1 servers and powers autonomous vehicle software and internal AI R&D across NVIDIA research, graphics, HPC, and robotics. This white paper is divided into the following sections: AI Workflow and Sizing NVIDIA AI Software DGX POD Design DGX POD Installation and Management 2018 NVIDIA Corporation. ii

4 1. AI Workflow and Sizing A typical AI software development workflow follows the steps shown in the following figure: Raw Data Labeled Data Adjust Model Parameters Trained Models Optimized Models Figure 1. AI workflow The workflow is detailed as follows: 1. A data factory collects raw data and includes tools used to pre-process, index, label, and manage data. 2. AI models are trained with labeled data using a DL framework from the NVIDIA GPU Cloud (NGC) container repository running on DGX servers with Volta Tensor Core GPUs. 3. AI model testing and validation adjusts model parameters as needed and repeats training until the desired accuracy is reached. 4. AI model optimization for production deployment (inference) is completed using the NVIDIA TensorRT optimizing inference accelerator. Sizing DL training is highly dependent on data size and model complexity. A single DGX-1 can complete a training experiment on a wide variety of AI models in one day. For example, the autonomous vehicle software team at NVIDIA developing NVIDIA DriveNet uses a custom Resnet-18 backbone detection network with 960x480x3 image size and trains at 480 images per second on a DGX-1 server allowing training of 120 epochs with 300k images in 21 hours. Internal experience at NVIDIA has shown that five developers collaborating on the development of one AI model provides the optimal development time. Each developer typically works on two models in parallel thus the infrastructure needs to support ten model training experiments within the desired TAT (turn-around-time). A DGX POD with nine DGX-1 servers can provide one day TAT for model training for the five-developer workgroup. During schedule-critical times, multi-node scaling can reduce turnaround time from one day to four hours using eight DGX-1 servers. Once in production, additional DGX-1 servers will be necessary to support on-going model refinement and regression testing NVIDIA Corporation. 1

5 2. NVIDIA AI Software NVIDIA AI software running on the DGX POD provides a high-performance DL training environment for large scale multi-user AI software development teams. NVIDIA AI software includes the DGX operating system (DGX OS), cluster management and orchestration tools, NVIDIA libraries and frameworks, workload schedulers, and optimized containers from the NGC container registry. To provide additional functionality, the DGX POD management software includes third-party open-source tools recommended by NVIDIA which have been tested to work on DGX PODs with the NVIDIA AI software stack. Support for these tools can be obtained directly through third-party support structures. Figure 2. NVIDIA AI software The foundation of the NVIDIA AI software stack is the DGX OS, built on an optimized version of the Ubuntu Linux operating system and tuned specifically for the DGX hardware. The DGX OS software includes certified GPU drivers, a network software stack, pre-configured NFS caching, NVIDIA data center GPU management (DCGM) diagnostic tools, GPU-enabled container runtime, NVIDIA CUDA SDK, cudnn, NCCL and other NVIDIA libraries, and support for NVIDIA GPUDirect. The DGX OS software can be automatically re-installed on demand by the DGX POD management software. The management software layer of the DGX POD (Figure 3) is composed of various services running on the Kubernetes container orchestration framework for fault tolerance and high availability. Services are provided for network configuration (DHCP) and fully-automated DGX OS software provisioning over the network (PXE) NVIDIA Corporation. 2

6 Monitoring of the DGX POD utilizes Prometheus for server data collection and storage in a timeseries database. Cluster-wide alerts are configured with Alertmanager, and DGX POD metrics are displayed using the Grafana web interface. For sites required to operate in an air-gapped environment or needing additional on-premises services, a local container registry mirroring NGC containers, as well as Ubuntu and Python package mirrors, can be run on the Kubernetes management layer to provide services to the cluster. Figure 3. DGX POD management software The DGX POD software allows for dynamic partitioning between the nodes assigned to Kubernetes and Slurm such that resources can be shifted between the partitions to meet the current workload demand. A simple user interface allows administrators to move DGX-1 servers between Kubernetes- and Slurm-managed domains. Kubernetes serves the dual role of running management services on management nodes as well as accepting user-defined workloads, and is installed on every server in the DGX POD. Slurm runs only user workloads and is installed on the login node as well as the DGX compute nodes. The DGX POD software allows individual DGX-1 servers to run jobs in either Kubernetes or Slurm. Kubernetes provides a high level of flexibility in load balancing, node failover, and bursting to external Kubernetes clusters, including NGC public cloud instances. Slurm supports a more static cluster environment but provides advanced HPC-style batch scheduling features including multinode scheduling that some workgroups may require. With the DGX POD software, idle systems can be moved back and forth as needed between the Kubernetes and Slurm environments. Future enhancements to Kubernetes are expected to support all DGX POD use cases in a pure Kubernetes environment. User workloads on the DGX POD primarily utilize containers from NGC (Figure 4), which provides researchers and data scientists with easy access to a comprehensive catalog of GPU-optimized software for DL, HPC applications, and HPC visualization that take full advantage of the GPUs. The NGC container registry includes NVIDIA tuned, tested, certified, and maintained containers for the top DL frameworks such as TensorFlow, PyTorch, and MXNet. NGC also has third-party managed HPC application containers, and NVIDIA HPC visualization containers NVIDIA Corporation. 3

7 Figure 4. NGC overview Management of the NVIDIA AI software on the DGX POD is accomplished with the Ansible configuration management tool. Ansible roles are used to install Kubernetes on the management nodes, install additional software on the login and DGX-1 servers, configure user accounts, configure external storage connections, install Kubernetes and Slurm schedulers, as well as performing day-to-day maintenance tasks such as new software installation, software updates, and GPU driver upgrades. The software management stack and documentation are available as an open source project on GitHub at: NVIDIA Corporation. 4

8 3. DGX POD The DGX POD (Figure 5) is an optimized data center rack containing up to nine DGX-1 servers, twelve storage servers, and three networking switches to support single and multi-node AI model training and inference using NVIDIA AI software. There are several factors to consider when planning a DGX POD deployment in order to determine if more than one rack is needed per DGX POD. This reference architecture is based on a single 35 kw high-density rack to provide the most efficient use of costly data center floorspace and to simplify network cabling. As GPU usage grows, the average power per server and power per rack continues to increase. However, older data centers may not yet be able to support the power and cooling densities required; hence the three-zone design allowing the DGX POD components to be installed in up to three lower-power racks. The DGX POD is designed to fit within a standard-height 42 RU data center rack. A taller rack can be used to include redundant networking switches, a management switch, and login servers. This reference architecture uses an additional utility rack for login and management servers, and has been sized and tested with up to six DGX PODs. Larger configurations of DGX PODs can be defined by an NVIDIA solution architect. A primary 10 GbE (minimum) network switch is used to connect all servers in the DGX POD and to provide access to a data center network. The DGX POD has been tested with an Arista switch with 48 x 10 GbE ports and 4 x 40 GbE uplinks. VLAN capabilities of the networking hardware are used to allow the out-of-band management network to run independently from the data network, while sharing the same physical hardware. Alternatively, a separate 1 GbE management switch may be used. While not included in the reference architecture, a second 10 GbE network switch can be used for redundancy and high availability. In addition to Arista, NVIDIA is working with other networking vendors who plan to release switch reference designs compatible with the DGX POD. A 36-port Mellanox 100 Gbps switch is configured to provide four 100 Gbps InfiniBand connections to the nine DGX-1 servers in the rack. This provides the best possible scalability for multi-node jobs. In the event of switch failure, multi-node jobs can fall back to use the 10 GbE switch for communications. The Mellanox switch can also be configured in 100 GbE mode for organizations that prefer to use Ethernet networking. Alternately, by configuring two 100 Gbps ports per DGX-1 server, the Mellanox switch can also be used by the storage servers. With the DGX family of servers, AI and HPC workloads are fusing into a unified architecture. For organizations that want to utilize multiple DGX PODs to run cluster-wide jobs, a core InfiniBand switch is configured in the utility rack in conjunction with a second 36-port Mellanox switch in DGX POD NVIDIA Corporation. 5

9 Nine DGX-1 servers (9 x 3 RU = 27 RU) Twelve storage servers (12 x 1 RU = 12 RU) 10 GbE (min) storage and management switch (1 RU) Mellanox 100 Gbps intra-rack high speed network switches (1 or 2 RU) Figure 5. Elevation of a DGX POD Storage architecture is important for optimized DL training performance. The DGX POD uses a hierarchical design with multiple levels of cache storage using the DGX-1 SSD and additional cache storage servers in the DGX POD. Long-term storage of raw data can be located on a wide variety of storage devices outside of the DGX POD, either on-premises or in public clouds. The DGX POD baseline storage architecture consists of standard NFS on the storage servers in conjunction with the local DGX SSD cache. Additional storage performance may be obtained by using the Ceph object-based file system or other caching file system on the storage servers NVIDIA Corporation. 6

10 The DGX POD is also designed to be compatible with a number of third-party storage solutions, see the reference architectures from DDN, NetApp, and Pure Storage for additional information. NVIDIA is also working with other storage vendors who plan to release DGX POD compatible reference architectures. While based on the DGX-1 server, the DGX POD has been designed in a modular fashion to support the NVIDIA DGX-2 server which starts shipping in Q3 of Each of the three compute zones in a DGX POD is designed such that the three DGX-1 servers in the compute zone can be replaced with a single DGX-2 server. A partial elevation of a DGX POD utility rack is shown in Figure 6. Login server which allows users to login to the cluster and launch Slurm batch jobs 1. (1 RU) Three management servers running Kubernetes server components and other DGX POD management software 2. (3 x 1 RU = 3 RU) Optional multi-pod 10 GbE storage and management network switches (2 RU) Optional multi-pod clustering using a Mellanox 216 port EDR InfiniBand switch. (12 RU) Figure 6. Partial elevation of a DGX POD utility rack 1 To support many users, the login server should have at two high-end CPUs, at least 1 TB of memory, two links to the 100 Gbps network, and redundant fans and power supplies. 2 These servers can be lower performance than the login server and can be configured with mid-range CPUs and less memory (128 to 256 GB) NVIDIA Corporation. 7

11 4. DGX POD Installation and Management Deploying a DGX POD is similar to deploying traditional servers and networking in a rack. However, with high-power consumption and corresponding cooling needs, server weight, and multiple networking cables per server, additional care and preparation is needed for a successful deployment. As with all IT equipment installation, it is important to work with the data center facilities team to ensure the DGX POD environmental requirements can be met. Additional DGX site requirements are detailed in the NVIDIA DGX Site Preparation Guide but important items to consider include: Area Rack Cooling 3 Power Design Guidelines Supports 3000 lbs of static load Dimensions of 1200 mm depth x 700 mm width Structured cabling pathways per TIA 942 standard Removal of 119,420 BTU/hr ASHRAE TC Thermal Guidelines Allowable Range North America: A/B power feeds, each three-phase 400V/60A/33.2kW (or three-phase 208V/60A/17.3 kw with additional considerations for redundancy as required) International: A/B power feeds, each 380/400/415V, 32A, three-phase 21-23kW each. 3 Via rack cooling door or data center hot/cold aisle air containment Table 1. Rack, cooling, and power considerations for a 35 kw DGX POD Figure 7 shows the server components and networking of the DGX POD. Management servers, login servers, DGX compute servers, and storage communicate over a 1 or 10 Gbps Ethernet network, while login servers, DGX compute servers and optionally storage can also communicate over high-speed 100 Gbps Infiniband or Ethernet. The DGX compute servers shown here are running both Kubernetes and Slurm to handle varying user workloads NVIDIA Corporation. 8

12 DGX Nodes Login Server Slurm Managed Kubernetes Managed Storage Servers WAN Management Nodes Figure 7. DGX POD networking Whether deploying single or multiple DGX PODs, it is best to use tools that take care of the management of all of the nodes. Use the NVIDIA AI software stack to provide an end-to-end solution from server OS installation to user job management. Installing the NVIDIA AI software stack on DGX PODs requires meeting a basic set of hardware and software prerequisites, setting local configuration parameters, and deploying the cluster following step-by-step instructions. Optional installation steps include configuring multiple job schedulers and integrating with enterprise authentication and storage systems. Once the PODs are powered-on and configured, verify all the components are working correctly. This can be an involved process. During this operational checkout, the following should be verified: Compute hardware, networking, and storage are operating as expected Power distribution works properly under maximum possible load Additional site tests as may be required for the data center The first step to validate DGX-1 server installation is performed using the diagnostic feature of DCGM. This tests many different GPU functions including memory, PCIe bus, SM units, memory bandwidth, and NVLink. DCGM will draw close to maximum server power during operation and 2018 NVIDIA Corporation. 9

13 thus is a good stress test of server power and cooling. By running this test simultaneously across all nodes using Kubernetes, it will stress the rack-level power, cooling, and airflow. The complete DCGM test takes approximately ten minutes to run on each DGX-1. This test should be run back to back for several hours to stress-test the DGX POD and ensure correct completion of each iteration. To test DL training, the NGC containers have built-in example scripts to check several different families of networks including image classification and language modeling via LSTMs. For example, the NGC TensorFlow Container contains scripts to test the major ImageNet Networks (Resnet-50, Inception, VGG-16, etc.) and allows for scalable server testing with real data or in synthetic mode. In addition, test any local applications that are planned to be run on the DGX POD. All performance results measured during server installation should be saved and used during routine retesting of the DGX POD to verify performance consistency as servers are modified and updated. Day-to-day operation and ongoing maintenance of a DGX POD is greatly simplified by the NVIDIA AI software stack. Typical operations such as installing new software, performing server upgrades, and managing scheduler allocations and reservations can all be handled automatically with simple commands NVIDIA Corporation. 10

14 5. Summary The DGX POD reference architecture provides organizations a blueprint to simplify deployment of GPU computing infrastructure to support large-scale AI software development efforts. A single DGX POD supporting small workgroups of AI developers can be grown into an infrastructure supporting thousands of users. The DGX POD reference architecture is based on the NVIDIA DGX SATURNV AI supercomputer which has 1000 DGX-1 servers and powers autonomous vehicle software and internal AI R&D across NVIDIA research, graphics, HPC, and robotics. The DGX POD has been design and tested by using specific storage and networking partners. In addition to those mentioned in this paper, NVIDIA is working with additional storage and networking vendors who plan to publish DGX POD-compatible reference architectures using their specific products. Because most installations of a DGX POD will require small differences such as cable length to integrate into your data center, NVIDIA does not sell DGX POD as a single unit. Work with an authorized NVIDIA Partner Network (NPN) reseller to configure and purchase a DGX POD. Finally, this white paper is meant to be a high-level overview and is not intended to be a step-by-step installation guide. Customers should work with an NPN provider to customize an installation plan for their organization NVIDIA Corporation. 11

15 Legal Notices and Trademarks Legal Notices ALL INFORMATION PROVIDED IN THIS WHITE PAPER, INCLUDING COMMENTARY, OPINION, NVIDIA DESIGN SPECIFICATIONS, REFERENCE BOARDS, FILES, DRAWINGS, DIAGNOSTICS, LISTS, AND OTHER DOCUMENTS (TOGETHER AND SEPARATELY, "MATERIALS") ARE BEING PROVIDED "AS IS." NVIDIA MAKES NO WARRANTIES, EXPRESSED, IMPLIED, STATUTORY, OR OTHERWISE WITH RESPECT TO MATERIALS, AND EXPRESSLY DISCLAIMS ALL IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTABILITY, AND FITNESS FOR A PARTICULAR PURPOSE. NVIDIA reserves the right to make corrections, modifications, enhancements, improvements, and other changes to this specification, at any time and/or to discontinue any product or service without notice. Customer should obtain the latest relevant specification before placing orders and should verify that such information is current and complete. NVIDIA products are sold subject to the NVIDIA standard terms and conditions of sale supplied at the time of order acknowledgement, unless otherwise agreed in an individual sales agreement signed by authorized representatives of NVIDIA and customer. NVIDIA hereby expressly objects to applying any customer general terms and conditions with regard to the purchase of the NVIDIA product referenced in this specification. NVIDIA products are not designed, authorized or warranted to be suitable for use in medical, military, aircraft, space or life support equipment, nor in applications where failure or malfunction of the NVIDIA product can reasonably be expected to result in personal injury, death or property or environmental damage. NVIDIA accepts no liability for inclusion and/or use of NVIDIA products in such equipment or applications and therefore such inclusion and/or use is at customer's own risk. NVIDIA makes no representation or warranty that products based on these specifications will be suitable for any specified use without further testing or modification. Testing of all parameters of each product is not necessarily performed by NVIDIA. It is customer's sole responsibility to ensure the product is suitable and fit for the application planned by customer and to do the necessary testing for the application in order to avoid a default of the application or the product. Weaknesses in customer's product designs may affect the quality and reliability of the NVIDIA product and may result in additional or different conditions and/or requirements beyond those contained in this specification. NVIDIA does not accept any liability related to any default, damage, costs or problem which may be based on or attributable to: (i) the use of the NVIDIA product in any manner that is contrary to this specification, or (ii) customer product designs. No license, either expressed or implied, is granted under any NVIDIA patent right, copyright, or other NVIDIA intellectual property right under this specification. Information published by NVIDIA regarding third-party products or services does not constitute a license from NVIDIA to use such products or services or a warranty or endorsement thereof. Use of such information may require a license from a third-party under the patents or other intellectual property rights of the third party, or a license from NVIDIA under the patents or other intellectual property rights of NVIDIA. Reproduction of information in this specification is permissible only if reproduction is approved by NVIDIA in writing, is reproduced without alteration, and is accompanied by all associated conditions, limitations, and notices. Trademarks NVIDIA, the NVIDIA logo, CUDA, Tesla, NVLink, DGX, DGX-1, and GPUDirect are trademarks or registered trademarks of NVIDIA Corporation in the United States and other countries. Other company and product names may be trademarks of the respective companies with which they are associated. Copyright 2018 NVIDIA Corporation. All rights reserved NVIDIA Corporation. iii

NVIDIA T4 FOR VIRTUALIZATION

NVIDIA T4 FOR VIRTUALIZATION NVIDIA T4 FOR VIRTUALIZATION TB-09377-001-v01 January 2019 Technical Brief TB-09377-001-v01 TABLE OF CONTENTS Powering Any Virtual Workload... 1 High-Performance Quadro Virtual Workstations... 3 Deep Learning

More information

MOSAIC CONTROL DISPLAYS

MOSAIC CONTROL DISPLAYS MOSAIC CONTROL DISPLAYS DA-06849-001_v02 May 2013 Application Note DOCUMENT CHANGE HISTORY DA-06849-001_v02 Version Date Authors Description of Change 01 May 1, 2013 AP, SM Initial Release 02 May 3, 2013

More information

NVIDIA GPU BOOST FOR TESLA

NVIDIA GPU BOOST FOR TESLA NVIDIA GPU BOOST FOR TESLA DA-06767-001_v02 January 2014 Application Note DOCUMENT CHANGE HISTORY DA-06767-001_v02 Version Date Authors Description of Change 01 March 28, 2013 GG, SM Initial Release 02

More information

NVIDIA DATA LOADING LIBRARY (DALI)

NVIDIA DATA LOADING LIBRARY (DALI) NVIDIA DATA LOADING LIBRARY (DALI) RN-09096-001 _v01 September 2018 Release Notes TABLE OF CONTENTS Chapter Chapter Chapter Chapter Chapter 1. 2. 3. 4. 5. DALI DALI DALI DALI DALI Overview...1 Release

More information

CREATING AN NVIDIA QUADRO VIRTUAL WORKSTATION INSTANCE

CREATING AN NVIDIA QUADRO VIRTUAL WORKSTATION INSTANCE CREATING AN NVIDIA QUADRO VIRTUAL WORKSTATION INSTANCE MICROSOFT AZURE MARKETPLACE DU-09260-001_v01 November 2018 Quick Start Guide DOCUMENT CHANGE HISTORY DU-09260-001_v01 Version Date Authors Description

More information

NVIDIA GPU CLOUD. DU _v02 July Getting Started Guide

NVIDIA GPU CLOUD. DU _v02 July Getting Started Guide NVIDIA GPU CLOUD DU-08787-001 _v02 July 2018 Getting Started Guide TABLE OF CONTENTS Chapter 1. Getting Started Using... 1 1.1. Signing Up for an NGC Account... 1 1.2. Logging on to the NGC Website...

More information

NVIDIA GPU CLOUD IMAGE FOR MICROSOFT AZURE

NVIDIA GPU CLOUD IMAGE FOR MICROSOFT AZURE NVIDIA GPU CLOUD IMAGE FOR MICROSOFT AZURE RN-08963-18.09.1 _v01 December 2018 Release Notes TABLE OF CONTENTS Chapter Chapter Chapter Chapter 1. 2. 3. 4. NVIDIA GPU Cloud Image Overview... 1 Version 18.11.1...

More information

NVIDIA GPU CLOUD IMAGE FOR GOOGLE CLOUD PLATFORM

NVIDIA GPU CLOUD IMAGE FOR GOOGLE CLOUD PLATFORM NVIDIA GPU CLOUD IMAGE FOR GOOGLE CLOUD PLATFORM RN-08963-18.06.0 _v01 June 2018 Release Notes TABLE OF CONTENTS 1. 2. 3. 4. 5. NVIDIA GPU Cloud Image Overview... 1 Version 18.06.0... 2 Version 18.05.0...

More information

NVIDIA VIRTUAL GPU PACKAGING, PRICING AND LICENSING. August 2017

NVIDIA VIRTUAL GPU PACKAGING, PRICING AND LICENSING. August 2017 NVIDIA VIRTUAL GPU PACKAGING, PRICING AND LICENSING August 2017 TABLE OF CONTENTS OVERVIEW... 3 GENERAL PURCHASING INFORMATION... 4 GRID PRODUCT DETAILS... 5 1.1 NVIDIA VIRTUAL GPU SOFTWARE EDITIONS...

More information

NVIDIA VIRTUAL GPU PACKAGING, PRICING AND LICENSING. March 2018 v2

NVIDIA VIRTUAL GPU PACKAGING, PRICING AND LICENSING. March 2018 v2 NVIDIA VIRTUAL GPU PACKAGING, PRICING AND LICENSING March 2018 v2 TABLE OF CONTENTS OVERVIEW 1.1 GENERAL PURCHASING INFORMATION VGPU PRODUCT DETAILS 1.2 VIRTUAL GPU SOFTWARE EDITIONS 1.3 VIRTUAL GPU SOFTWARE

More information

NVIDIA VOLTA DEEP LEARNING AMI

NVIDIA VOLTA DEEP LEARNING AMI NVIDIA VOLTA DEEP LEARNING AMI RN-08791-18.09.1 _v01 September 2018 Release Notes TABLE OF CONTENTS 1. Overview... 1 2. Version 18.09.1... 2 3. Version 18.08.0... 3 4. Version 18.07.0... 4 5. Version 18.06.0...

More information

TESLA K20X GPU ACCELERATOR

TESLA K20X GPU ACCELERATOR TESLA K20 GPU ACCELERATOR BD-06397-001_v09 October 2014 Board Specification DOCUMENT CHANGE HISTORY BD-06397-001_v09 Version Date Authors Description of Change 01 July 31, 2012 GG, SM Preliminary Information

More information

TESLA P6. PB _v02 August Product Brief

TESLA P6. PB _v02 August Product Brief TESLA P6 PB-08482-001_v02 August 2017 Product Brief DOCUMENT CHANGE HISTORY PB-08482-001_v02 Version Date Authors Description of Change 01 March 24, 2017 VK, DV Initial release 02 August 31, 2017 VK, DV

More information

DGX SOFTWARE FOR RED HAT ENTERPRISE LINUX 7

DGX SOFTWARE FOR RED HAT ENTERPRISE LINUX 7 DGX SOFTWARE FOR RED HAT ENTERPRISE LINUX 7 RN-09325-001 _v18.11 December 2018 Release Notes TABLE OF CONTENTS Chapter 1. DGX Software For Red Hat Enterprise Linux 7 Overview...1 Chapter 2. Version EL7-18.11...

More information

TESLA V100 PCIe GPU ACCELERATOR

TESLA V100 PCIe GPU ACCELERATOR TESLA V100 PCIe GPU ACCELERATOR PB-08744-001_v03 October 2017 Product Brief DOCUMENT CHANGE HISTORY PB-08744-001_v03 Version Date Authors Description of Change 01 September 19, 2017 GG, SM Initial Release

More information

BEST PRACTICES FOR DOCKER

BEST PRACTICES FOR DOCKER BEST PRACTICES FOR DOCKER DG-08863-001 _v001 December 2018 Best Practices TABLE OF CONTENTS Chapter 1. NVIDIA Container Best Practices...1 1.1. Hello World For Containers... 1 1.2. Logging Into Docker...

More information

NVIDIA DGX SYSTEMS PURPOSE-BUILT FOR AI

NVIDIA DGX SYSTEMS PURPOSE-BUILT FOR AI NVIDIA DGX SYSTEMS PURPOSE-BUILT FOR AI Overview Unparalleled Value Product Portfolio Software Platform From Desk to Data Center to Cloud Summary AI researchers depend on computing performance to gain

More information

CUDNN. DU _v07 December Installation Guide

CUDNN. DU _v07 December Installation Guide CUDNN DU-08670-001_v07 December 2017 Installation Guide TABLE OF CONTENTS Chapter Overview... 1 Chapter Installing on Linux... 2 Prerequisites... 2 Installing NVIDIA Graphics Drivers... 2 Installing CUDA...

More information

DGX UPDATE. Customer Presentation Deck May 8, 2017

DGX UPDATE. Customer Presentation Deck May 8, 2017 DGX UPDATE Customer Presentation Deck May 8, 2017 NVIDIA DGX-1: The World s Fastest AI Supercomputer FASTEST PATH TO DEEP LEARNING EFFORTLESS PRODUCTIVITY REVOLUTIONARY AI PERFORMANCE Fully-integrated

More information

PREPARING TO USE CONTAINERS

PREPARING TO USE CONTAINERS PREPARING TO USE CONTAINERS DU-08786-001_v001 May 2018 Getting Started Guide TABLE OF CONTENTS Chapter Introduction To Docker And Containers...1 Chapter 2. Preparing Your DGX System For Use With nvidia-docker...

More information

TESLA P4 GPU ACCELERATOR

TESLA P4 GPU ACCELERATOR TESLA P4 GPU ACCELERATOR PB-08449-001_v01 February 2017 Product Brief DOCUMENT CHANGE HISTORY PB-08449-001_v01 Version Date Authors Description of Change 01 February 15, 2017 HV, SM Initial Release Tesla

More information

NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL)

NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL) NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL) RN-08645-000_v01 September 2018 Release Notes TABLE OF CONTENTS Chapter Chapter Chapter Chapter Chapter Chapter Chapter Chapter Chapter Chapter 1. NCCL Overview...1

More information

BEST PRACTICES FOR DOCKER

BEST PRACTICES FOR DOCKER BEST PRACTICES FOR DOCKER DG-08863-001 _v001 March 2018 Best Practices TABLE OF CONTENTS Chapter 1. Docker Best Practices with NVIDIA Containers... 1 1.1. Prerequisites... 1 1.1.1. Hello World For Containers...

More information

NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL)

NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL) NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL) RN-08645-000_v01 March 2018 Release Notes TABLE OF CONTENTS Chapter Chapter Chapter Chapter Chapter Chapter Chapter 1. 2. 3. 4. 5. 6. 7. NCCL NCCL NCCL NCCL

More information

IBM SpectrumAI with NVIDIA Converged Infrastructure Solutions for AI workloads

IBM SpectrumAI with NVIDIA Converged Infrastructure Solutions for AI workloads IBM SpectrumAI with NVIDIA Converged Infrastructure Solutions for AI workloads The engine to power your AI data pipeline Introduction: Artificial intelligence (AI) including deep learning (DL) and machine

More information

TENSORRT 4.0 RELEASE CANDIDATE (RC)

TENSORRT 4.0 RELEASE CANDIDATE (RC) TENSORRT 4.0 RELEASE CANDIDATE (RC) DU-08731-001_v4.0 RC March 2018 Installation Guide TABLE OF CONTENTS Chapter 1. Overview... 1 Chapter 2. Getting Started... 2 Chapter 3. Downloading TensorRT...3 Chapter

More information

TENSORRT 3.0. DU _v3.0 February Installation Guide

TENSORRT 3.0. DU _v3.0 February Installation Guide TENSORRT 3.0 DU-08731-001_v3.0 February 2018 Installation Guide TABLE OF CONTENTS Chapter 1. Overview... 1 Chapter 2. Getting Started... 2 Chapter 3. Downloading TensorRT...4 Chapter 4. Installing TensorRT...

More information

NVIDIA GRID K1 AND NVIDIA GRID K2 CERTIFIED OEM PLATFORMS

NVIDIA GRID K1 AND NVIDIA GRID K2 CERTIFIED OEM PLATFORMS K1 AND NVIDIA GRID K2 CERTIFIED OEM PLATFORMS DA-09018-001_v01 May 2018 Application Note DOCUMENT CHANGE HISTORY DA-09018-001_v01 Version Date Authors Description of Change 01 May 17, 2018 VNK, SM Initial

More information

NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL)

NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL) NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL) DU-08730-210_v01 March 2018 Installation Guide TABLE OF CONTENTS Chapter 1. Overview... 1 Chapter 2. Prerequisites...3 2.1. Software Requirements... 3 2.2.

More information

IBM Power AC922 Server

IBM Power AC922 Server IBM Power AC922 Server The Best Server for Enterprise AI Highlights More accuracy - GPUs access system RAM for larger models Faster insights - significant deep learning speedups Rapid deployment - integrated

More information

TESLA P40 GPU ACCELERATOR

TESLA P40 GPU ACCELERATOR TESLA P40 GPU ACCELERATOR PB-08338-001_v01 November 2016 Product Brief DOCUMENT CHANGE HISTORY PB-08338-001_v01 Version Date Authors Description of Change 01 November 29, 2016 WT, SM Initial Release Tesla

More information

NVIDIA GPU CLOUD DEEP LEARNING FRAMEWORKS

NVIDIA GPU CLOUD DEEP LEARNING FRAMEWORKS TECHNICAL OVERVIEW NVIDIA GPU CLOUD DEEP LEARNING FRAMEWORKS A Guide to the Optimized Framework Containers on NVIDIA GPU Cloud Introduction Artificial intelligence is helping to solve some of the most

More information

IBM Power Advanced Compute (AC) AC922 Server

IBM Power Advanced Compute (AC) AC922 Server IBM Power Advanced Compute (AC) AC922 Server The Best Server for Enterprise AI Highlights IBM Power Systems Accelerated Compute (AC922) server is an acceleration superhighway to enterprise- class AI. A

More information

USING NGC WITH YOUR NVIDIA TITAN PC

USING NGC WITH YOUR NVIDIA TITAN PC USING NGC WITH YOUR NVIDIA TITAN PC DU-08832-001_v02 December 2017 Setup Guide TABLE OF CONTENTS Chapter 1. Introduction...1 Chapter 2. Installing the NVIDIA Driver... 2 2.1. Setting Up the Driver Repository...

More information

IBM Power Systems HPC Cluster

IBM Power Systems HPC Cluster IBM Power Systems HPC Cluster Highlights Complete and fully Integrated HPC cluster for demanding workloads Modular and Extensible: match components & configurations to meet demands Integrated: racked &

More information

Intel Cluster Ready Allowed Hardware Variances

Intel Cluster Ready Allowed Hardware Variances Intel Cluster Ready Allowed Hardware Variances Solution designs are certified as Intel Cluster Ready with an exact bill of materials for the hardware and the software stack. When instances of the certified

More information

SYNERGIE VON HPC UND DEEP LEARNING MIT NVIDIA GPUS

SYNERGIE VON HPC UND DEEP LEARNING MIT NVIDIA GPUS SYNERGIE VON HPC UND DEEP LEARNING MIT NVIDIA S Axel Koehler, Principal Solution Architect HPCN%Workshop%Goettingen,%14.%Mai%2018 NVIDIA - AI COMPUTING COMPANY Computer Graphics Computing Artificial Intelligence

More information

S INSIDE NVIDIA GPU CLOUD DEEP LEARNING FRAMEWORK CONTAINERS

S INSIDE NVIDIA GPU CLOUD DEEP LEARNING FRAMEWORK CONTAINERS S8497 - INSIDE NVIDIA GPU CLOUD DEEP LEARNING FRAMEWORK CONTAINERS Chris Lamb CUDA and NGC Engineering, NVIDIA John Barco NGC Product Management, NVIDIA NVIDIA GPU Cloud (NGC) overview AGENDA Using NGC

More information

TESLA M10. PB _v04 February Product Brief

TESLA M10. PB _v04 February Product Brief TESLA M10 PB-08118-001_v04 February 2017 Product Brief DOCUMENT CHANGE HISTORY PB-08118-001_v04 Version Date Authors Description of Change 01 June 17, 2016 MD, SM Initial Release 02 October 4, 2016 MD,

More information

TESLA DRIVER VERSION (LINUX)/411.98(WINDOWS)

TESLA DRIVER VERSION (LINUX)/411.98(WINDOWS) TESLA DRIVER VERSION 410.79(LINUX)/411.98(WINDOWS) RN-08625-410.79_411.98 _v01 December 2018 Release Notes TABLE OF CONTENTS Chapter 1. Version Highlights...1 1.1. Fixed Issues... 1 1.2. Known Issues...1

More information

S THE MAKING OF DGX SATURNV: BREAKING THE BARRIERS TO AI SCALE. Presenter: Louis Capps, Solution Architect, NVIDIA,

S THE MAKING OF DGX SATURNV: BREAKING THE BARRIERS TO AI SCALE. Presenter: Louis Capps, Solution Architect, NVIDIA, S7750 - THE MAKING OF DGX SATURNV: BREAKING THE BARRIERS TO AI SCALE Presenter: Louis Capps, Solution Architect, NVIDIA, lcapps@nvidia.com A TALE OF ENLIGHTENMENT Basic OK List 10 for x = 1 to 3 20 print

More information

TESLA V100 PERFORMANCE GUIDE. Life Sciences Applications

TESLA V100 PERFORMANCE GUIDE. Life Sciences Applications TESLA V100 PERFORMANCE GUIDE Life Sciences Applications NOVEMBER 2017 TESLA V100 PERFORMANCE GUIDE Modern high performance computing (HPC) data centers are key to solving some of the world s most important

More information

Network Service Description

Network Service Description Network Service Description Applies to: Office 365 Dedicated Topic Last Modified: 2015-09-03 Contents... 1 Network Architecture... 2 Customer Connectivity to Services... 5 Customer-Owned Private Network

More information

CS500 SMARTER CLUSTER SUPERCOMPUTERS

CS500 SMARTER CLUSTER SUPERCOMPUTERS CS500 SMARTER CLUSTER SUPERCOMPUTERS OVERVIEW Extending the boundaries of what you can achieve takes reliable computing tools matched to your workloads. That s why we tailor the Cray CS500 cluster supercomputer

More information

Scalable AI Infrastructure

Scalable AI Infrastructure Technical Whitepaper Scalable AI Infrastructure Designing For Real-World Deep Learning Use Cases Sundar Ranganathan, NetApp Santosh Rao, NetApp June 2018 WP-7267 In partnership with Executive Summary Deep

More information

VRWORKS AUDIO SDK. DU _v01 May Overview

VRWORKS AUDIO SDK. DU _v01 May Overview VRWORKS AUDIO SDK DU-08562-001_v01 May 2017 Overview DOCUMENT CHANGE HISTORY DU-08562-001_v01 Version Date Authors Description of Change 01 May 2017 AS Initial release VRWorks Audio SDK DU-08562-001_v01

More information

TESLA DRIVER VERSION (LINUX)/411.82(WINDOWS)

TESLA DRIVER VERSION (LINUX)/411.82(WINDOWS) TESLA DRIVER VERSION 410.72(LINUX)/411.82(WINDOWS) RN-08625-410.72_411.82 _v01 October 2018 Release Notes TABLE OF CONTENTS Chapter 1. Version Highlights...1 1.1. New Features...1 1.2. Improvements...

More information

NVIDIA DIGITS CONTAINER

NVIDIA DIGITS CONTAINER NVIDIA DIGITS CONTAINER DU-09194-001 _v1.0 January 2019 User Guide TABLE OF CONTENTS Chapter 1. Overview... 1 Chapter 2. Creating A Dataset Using Data From An S3 Endpoint... 2 Chapter 3. Writing a DIGITS

More information

Cisco UCS C480 ML M5 Rack Server Performance Characterization

Cisco UCS C480 ML M5 Rack Server Performance Characterization White Paper Cisco UCS C480 ML M5 Rack Server Performance Characterization The Cisco UCS C480 ML M5 Rack Server platform is designed for artificial intelligence and machine-learning workloads. 2018 Cisco

More information

IBM Deep Learning Solutions

IBM Deep Learning Solutions IBM Deep Learning Solutions Reference Architecture for Deep Learning on POWER8, P100, and NVLink October, 2016 How do you teach a computer to Perceive? 2 Deep Learning: teaching Siri to recognize a bicycle

More information

TESLA M60. PB _v02 December Product Brief

TESLA M60. PB _v02 December Product Brief TESLA M60 PB-07864-001_v02 December 2016 Product Brief DOCUMENT CHANGE HISTORY PB-07864-001_v02 Version Date Authors Description of Change 01 November 13, 2015 MD, SM Initial Release 02 December 9, 2016

More information

SUN CUSTOMER READY HPC CLUSTER: REFERENCE CONFIGURATIONS WITH SUN FIRE X4100, X4200, AND X4600 SERVERS Jeff Lu, Systems Group Sun BluePrints OnLine

SUN CUSTOMER READY HPC CLUSTER: REFERENCE CONFIGURATIONS WITH SUN FIRE X4100, X4200, AND X4600 SERVERS Jeff Lu, Systems Group Sun BluePrints OnLine SUN CUSTOMER READY HPC CLUSTER: REFERENCE CONFIGURATIONS WITH SUN FIRE X4100, X4200, AND X4600 SERVERS Jeff Lu, Systems Group Sun BluePrints OnLine April 2007 Part No 820-1270-11 Revision 1.1, 4/18/07

More information

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND FIBRE CHANNEL INFRASTRUCTURE

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND FIBRE CHANNEL INFRASTRUCTURE DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND FIBRE CHANNEL INFRASTRUCTURE Design Guide APRIL 0 The information in this publication is provided as is. Dell Inc. makes no representations or warranties

More information

SUSE OpenStack Cloud Production Deployment Architecture. Guide. Solution Guide Cloud Computing.

SUSE OpenStack Cloud Production Deployment Architecture. Guide. Solution Guide Cloud Computing. SUSE OpenStack Cloud Production Deployment Architecture Guide Solution Guide Cloud Computing Table of Contents page Introduction... 2 High Availability Configuration...6 Network Topography...8 Services

More information

Democratizing Machine Learning on Kubernetes

Democratizing Machine Learning on Kubernetes Democratizing Machine Learning on Kubernetes Joy Qiao, Senior Solution Architect - AI and Research Group, Microsoft Lachlan Evenson - Principal Program Manager AKS/ACS, Microsoft Who are we? The Data Scientist

More information

S8765 Performance Optimization for Deep- Learning on the Latest POWER Systems

S8765 Performance Optimization for Deep- Learning on the Latest POWER Systems S8765 Performance Optimization for Deep- Learning on the Latest POWER Systems Khoa Huynh Senior Technical Staff Member (STSM), IBM Jonathan Samn Software Engineer, IBM Evolving from compute systems to

More information

DGX SYSTEMS: DEEP LEARNING FROM DESK TO DATA CENTER. Markus Weber and Haiduong Vo

DGX SYSTEMS: DEEP LEARNING FROM DESK TO DATA CENTER. Markus Weber and Haiduong Vo DGX SYSTEMS: DEEP LEARNING FROM DESK TO DATA CENTER Markus Weber and Haiduong Vo NVIDIA DGX SYSTEMS Agenda NVIDIA DGX-1 NVIDIA DGX STATION 2 ONE YEAR LATER NVIDIA DGX-1 Barriers Toppled, the Unsolvable

More information

AIRI SCALE-OUT AI-READY INFRASTRUCTURE ARCHITECTED BY PURE STORAGE AND NVIDIA WITH ARISTA 7060X SWITCH REFERENCE ARCHITECTURE

AIRI SCALE-OUT AI-READY INFRASTRUCTURE ARCHITECTED BY PURE STORAGE AND NVIDIA WITH ARISTA 7060X SWITCH REFERENCE ARCHITECTURE REFERENCE ARCHITECTURE AIRI SCALE-OUT AI-READY INFRASTRUCTURE ARCHITECTED BY PURE STORAGE AND NVIDIA WITH ARISTA 7060X SWITCH TABLE OF CONTENTS INTRODUCTION... 3 Accelerating Computation: NVIDIA DGX-1...

More information

QuickSpecs. Models. HP NC380T PCI Express Dual Port Multifunction Gigabit Server Adapter. Overview

QuickSpecs. Models. HP NC380T PCI Express Dual Port Multifunction Gigabit Server Adapter. Overview Overview The HP NC380T server adapter is the industry's first PCI Express dual port multifunction network adapter supporting TOE (TCP/IP Offload Engine) for Windows, iscsi (Internet Small Computer System

More information

Intel Cloud Builder Guide: Cloud Design and Deployment on Intel Platforms

Intel Cloud Builder Guide: Cloud Design and Deployment on Intel Platforms EXECUTIVE SUMMARY Intel Cloud Builder Guide Intel Xeon Processor-based Servers Novell* Cloud Manager Intel Cloud Builder Guide: Cloud Design and Deployment on Intel Platforms Novell* Cloud Manager Intel

More information

DGX SOFTWARE WITH RED HAT ENTERPRISE LINUX 7

DGX SOFTWARE WITH RED HAT ENTERPRISE LINUX 7 DGX SOFTWARE WITH RED HAT ENTERPRISE LINUX 7 RN-09301-001 _v02 January 2019 Installation Guide TABLE OF CONTENTS Chapter 1. Introduction...1 1.1. Related Documentation... 1 1.2. Prerequisites... 1 1.2.1.

More information

Oracle Database Appliance X6-2L

Oracle Database Appliance X6-2L Oracle Database Appliance X6-2L The Oracle Database Appliance saves time and money by simplifying deployment, maintenance, and support of database solutions for organizations of every size. Optimized for

More information

IBM TotalStorage SAN Switch F32

IBM TotalStorage SAN Switch F32 Intelligent fabric switch with enterprise performance for midrange and large storage networks IBM TotalStorage SAN Switch F32 High port density packaging helps save rack space Highlights Can be used as

More information

HPE ProLiant ML350 Gen10 Server

HPE ProLiant ML350 Gen10 Server Digital data sheet HPE ProLiant ML350 Gen10 Server ProLiant ML Servers What's new Support for Intel Xeon Scalable processors full stack. 2600 MT/s HPE DDR4 SmartMemory RDIMM/LRDIMM offering 8, 16, 32,

More information

Architecting High Performance Computing Systems for Fault Tolerance and Reliability

Architecting High Performance Computing Systems for Fault Tolerance and Reliability Architecting High Performance Computing Systems for Fault Tolerance and Reliability Blake T. Gonzales HPC Computer Scientist Dell Advanced Systems Group blake_gonzales@dell.com Agenda HPC Fault Tolerance

More information

Building the Most Efficient Machine Learning System

Building the Most Efficient Machine Learning System Building the Most Efficient Machine Learning System Mellanox The Artificial Intelligence Interconnect Company June 2017 Mellanox Overview Company Headquarters Yokneam, Israel Sunnyvale, California Worldwide

More information

DU _v01. September User Guide

DU _v01. September User Guide NVIDIA MAXIMUS TECHNOLOGY FOR ANSYS MECHANICAL DU-06467-001_v01 September 2012 User Guide DOCUMENT CHANGE HISTORY DU-06467-001_v01 Version Date Authors Description of Change 01 August 3, 2012 Initial release

More information

DGX-2 SYSTEM FIRMWARE UPDATE CONTAINER

DGX-2 SYSTEM FIRMWARE UPDATE CONTAINER DGX-2 SYSTEM FIRMWARE UPDATE CONTAINER RN-09214-001 _v18.10.2 November 2018 Release Notes TABLE OF CONTENTS Chapter 1. DGX-2 System FW Update Container Overview... 1 Chapter 2. Version 18.10.2... 2 2.1.

More information

SUN ORACLE DATABASE MACHINE

SUN ORACLE DATABASE MACHINE SUN ORACLE DATABASE MACHINE FEATURES AND FACTS FEATURES From 1 to 8 database servers From 1 to 14 Sun Oracle Exadata Storage Servers Each Exadata Storage Server includes 384 GB of Exadata QDR (40 Gb/second)

More information

Building the Most Efficient Machine Learning System

Building the Most Efficient Machine Learning System Building the Most Efficient Machine Learning System Mellanox The Artificial Intelligence Interconnect Company June 2017 Mellanox Overview Company Headquarters Yokneam, Israel Sunnyvale, California Worldwide

More information

Oracle Database Appliance X6-2S / X6-2M

Oracle Database Appliance X6-2S / X6-2M Oracle Database Appliance X6-2S / X6-2M The Oracle Database Appliance saves time and money by simplifying deployment, maintenance, and support of database solutions for organizations of every size. Optimized

More information

World s most advanced data center accelerator for PCIe-based servers

World s most advanced data center accelerator for PCIe-based servers NVIDIA TESLA P100 GPU ACCELERATOR World s most advanced data center accelerator for PCIe-based servers HPC data centers need to support the ever-growing demands of scientists and researchers while staying

More information

Preparing GPU-Accelerated Applications for the Summit Supercomputer

Preparing GPU-Accelerated Applications for the Summit Supercomputer Preparing GPU-Accelerated Applications for the Summit Supercomputer Fernanda Foertter HPC User Assistance Group Training Lead foertterfs@ornl.gov This research used resources of the Oak Ridge Leadership

More information

Sun Fire X4170 M2 Server Frequently Asked Questions

Sun Fire X4170 M2 Server Frequently Asked Questions Overview Faced with ever increasing computing needs and budget constraints, companies today want to set up infrastructures that offer optimal value, can easily be re-purposed, and have reduced complexity.

More information

UCS M-Series + Citrix XenApp Optimizing high density XenApp deployment at Scale

UCS M-Series + Citrix XenApp Optimizing high density XenApp deployment at Scale In Collaboration with Intel UCS M-Series + Citrix XenApp Optimizing high density XenApp deployment at Scale Aniket Patankar UCS Product Manager May 2015 Cisco UCS - Powering Applications at Every Scale

More information

An Oracle White Paper December Accelerating Deployment of Virtualized Infrastructures with the Oracle VM Blade Cluster Reference Configuration

An Oracle White Paper December Accelerating Deployment of Virtualized Infrastructures with the Oracle VM Blade Cluster Reference Configuration An Oracle White Paper December 2010 Accelerating Deployment of Virtualized Infrastructures with the Oracle VM Blade Cluster Reference Configuration Introduction...1 Overview of the Oracle VM Blade Cluster

More information

Super-convergence for Real Time Analytics at the Edge Flashmatrix 1000 Series. storage.toshiba.com/flashmatrix/

Super-convergence for Real Time Analytics at the Edge Flashmatrix 1000 Series. storage.toshiba.com/flashmatrix/ Super-convergence for Real Time Analytics at the Edge Flashmatrix 1000 Series About the Flashmatrix 1000 Series Edge computing represents the next significant wave in computing. The IoT is ushering in

More information

ORACLE DATABASE APPLIANCE X3 2

ORACLE DATABASE APPLIANCE X3 2 ORACLE DATABASE APPLIANCE X3 2 FEATURES Fully integrated database appliance Simple one button installation, patching, and diagnostics Oracle Database, Enterprise Edition Oracle Real Application Clusters

More information

IBM Z servers running Oracle Database 12c on Linux

IBM Z servers running Oracle Database 12c on Linux IBM Z servers running Oracle Database 12c on Linux Put Z to work for you Scale and grow Oracle Database 12c applications and data with confidence Benefit from mission-critical reliability for Oracle Database

More information

Introducing VMware Validated Design Use Cases. Modified on 21 DEC 2017 VMware Validated Design 4.1

Introducing VMware Validated Design Use Cases. Modified on 21 DEC 2017 VMware Validated Design 4.1 Introducing VMware Validated Design Use Cases Modified on 21 DEC 2017 VMware Validated Design 4.1 You can find the most up-to-date technical documentation on the VMware website at: https://docs.vmware.com/

More information

Deep Learning mit PowerAI - Ein Überblick

Deep Learning mit PowerAI - Ein Überblick Stephen Lutz Deep Learning mit PowerAI - Open Group Master Certified IT Specialist Technical Sales IBM Cognitive Infrastructure IBM Germany Ein Überblick Stephen.Lutz@de.ibm.com What s that? and what s

More information

New Approach to Unstructured Data

New Approach to Unstructured Data Innovations in All-Flash Storage Deliver a New Approach to Unstructured Data Table of Contents Developing a new approach to unstructured data...2 Designing a new storage architecture...2 Understanding

More information

Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance

Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance This Dell EMC technical white paper discusses performance benchmarking results and analysis for Simulia

More information

Data Center Management and Automation Strategic Briefing

Data Center Management and Automation Strategic Briefing Data Center and Automation Strategic Briefing Contents Why is Data Center and Automation (DCMA) so important? 2 The Solution Pathway: Data Center and Automation 2 Identifying and Addressing the Challenges

More information

TACKLING THE CHALLENGES OF NEXT GENERATION HEALTHCARE

TACKLING THE CHALLENGES OF NEXT GENERATION HEALTHCARE TACKLING THE CHALLENGES OF NEXT GENERATION HEALTHCARE Nicola Rieke, Senior Deep Learning Solution Architect Healthcare EMEA Fausto Milletari, Senior Deep Learning Solution Architect Healthcare NALA INTRODUCTION

More information

EMC SYMMETRIX VMAX 40K SYSTEM

EMC SYMMETRIX VMAX 40K SYSTEM EMC SYMMETRIX VMAX 40K SYSTEM The EMC Symmetrix VMAX 40K storage system delivers unmatched scalability and high availability for the enterprise while providing market-leading functionality to accelerate

More information

HP ProLiant m300 1P C2750 CPU 32GB Configure-to-order Server Cartridge B21

HP ProLiant m300 1P C2750 CPU 32GB Configure-to-order Server Cartridge B21 Overview HP Moonshot System with the HP ProLiant m300 server cartridge was created for cost-effective Dynamic Content Delivery, frontend Web and online analytics. The ProLiant m300 server cartridge has

More information

EMC SYMMETRIX VMAX 40K STORAGE SYSTEM

EMC SYMMETRIX VMAX 40K STORAGE SYSTEM EMC SYMMETRIX VMAX 40K STORAGE SYSTEM The EMC Symmetrix VMAX 40K storage system delivers unmatched scalability and high availability for the enterprise while providing market-leading functionality to accelerate

More information

HPE ProLiant ML110 Gen10 Server

HPE ProLiant ML110 Gen10 Server Digital data sheet HPE ProLiant ML110 Gen10 Server ProLiant ML Servers What's new New SMB focused offers regionally released as Smart Buy Express in the U.S. and Canada, Top Value in Europe, and Intelligent

More information

HPE ProLiant ML350 Gen P 16GB-R E208i-a 8SFF 1x800W RPS Solution Server (P04674-S01)

HPE ProLiant ML350 Gen P 16GB-R E208i-a 8SFF 1x800W RPS Solution Server (P04674-S01) Digital data sheet HPE ProLiant ML350 Gen10 4110 1P 16GB-R E208i-a 8SFF 1x800W RPS Solution Server (P04674-S01) ProLiant ML Servers What's new Support for Intel Xeon Scalable processors full stack. 2600

More information

HPE ProLiant DL360 Gen P 16GB-R P408i-a 8SFF 500W PS Performance Server (P06453-B21)

HPE ProLiant DL360 Gen P 16GB-R P408i-a 8SFF 500W PS Performance Server (P06453-B21) Digital data sheet HPE ProLiant DL360 Gen10 4110 1P 16GB-R P408i-a 8SFF 500W PS Performance Server (P06453-B21) ProLiant DL Servers What's new Innovative design with greater flexibility to mix and match

More information

ACCELERATED COMPUTING: THE PATH FORWARD. Jensen Huang, Founder & CEO SC17 Nov. 13, 2017

ACCELERATED COMPUTING: THE PATH FORWARD. Jensen Huang, Founder & CEO SC17 Nov. 13, 2017 ACCELERATED COMPUTING: THE PATH FORWARD Jensen Huang, Founder & CEO SC17 Nov. 13, 2017 COMPUTING AFTER MOORE S LAW Tech Walker 40 Years of CPU Trend Data 10 7 GPU-Accelerated Computing 10 5 1.1X per year

More information

HPE ProLiant DL580 Gen10 Server

HPE ProLiant DL580 Gen10 Server Digital data sheet HPE ProLiant DL580 Gen10 Server ProLiant DL Servers What's new Support for the new Intel Xeon Scalable Gold 6143 and Intel Xeon Scalable Platinum 8165 processors which support core boosting

More information

Copyright 2015 EMC Corporation. All rights reserved. Published in the USA.

Copyright 2015 EMC Corporation. All rights reserved. Published in the USA. This Reference Architecture Guide describes, in summary, a solution that enables IT organizations to quickly and effectively provision and manage Oracle Database as a Service (DBaaS) on Federation Enterprise

More information

HPE ProLiant ML110 Gen P 8GB-R S100i 4LFF NHP SATA 350W PS DVD Entry Server/TV (P )

HPE ProLiant ML110 Gen P 8GB-R S100i 4LFF NHP SATA 350W PS DVD Entry Server/TV (P ) Digital data sheet HPE ProLiant ML110 Gen10 3104 1P 8GB-R S100i 4LFF NHP SATA 350W PS DVD Entry Server/TV (P03684-425) ProLiant ML Servers What's new New SMB focused offers regionally released as Smart

More information

Avaya Collaboration Pod 4200 Series

Avaya Collaboration Pod 4200 Series Highlights Simplify your installation with a pre-integrated virtualized solution: VMware-based solution helps reduce infrastructure costs and time to deploy Faster time to service: Ready-made, pre-tested

More information

Deep Learning Inference on Openshift with GPUs

Deep Learning Inference on Openshift with GPUs Deep Learning Inference on Openshift with GPUs OpenShift Commons, Seattle, Dec 10 2018 Tripti Singhal Product Manager, NVIDIA Deep Learning Software Tushar Katarki Product Manager, AI on OpenShift AGENDA

More information

IBM System Storage DS4800

IBM System Storage DS4800 Scalable, high-performance storage for on demand computing environments IBM System Storage DS4800 Highlights 4 Gbps Fibre Channel interface Includes IBM System Storage technology DS4000 Storage Manager

More information

IBM System Storage DS6800

IBM System Storage DS6800 Enterprise-class storage in a small, scalable package IBM System Storage DS6800 Highlights Designed to deliver Designed to provide over enterprise-class functionality, 1600 MBps performance for with open

More information

EMC Symmetrix VMAX Cloud Edition

EMC Symmetrix VMAX Cloud Edition EMC Symmetrix VMAX Cloud Edition VERSION 2.2 Automation Appliance Physical Planning Guide REV 02 Copyright 2013-2013 EMC Corporation. All rights reserved. Published in the USA. Published December, 2013

More information