Introduction to Cloudbreak

Similar documents
Troubleshooting Cloudbreak

Accessing clusters 2. Accessing Clusters. Date of Publish:

Managing and Monitoring Cloudbreak

SAP Vora - AWS Marketplace Production Edition Reference Guide

HDP Security Overview

HDP Security Overview

Hortonworks Data Platform

Administration 1. DLM Administration. Date of Publish:

Configuring Apache Knox SSO

Cisco CloudCenter Solution with VMware

At Course Completion Prepares you as per certification requirements for AWS Developer Associate.

Installing Apache Zeppelin

Hortonworks Data Platform

SAP VORA 1.4 on AWS - MARKETPLACE EDITION FREQUENTLY ASKED QUESTIONS

Getting Started 1. Getting Started. Date of Publish:

New Features and Enhancements in Big Data Management 10.2

Configuring Apache Knox SSO

Cloudera s Enterprise Data Hub on the Amazon Web Services Cloud: Quick Start Reference Deployment October 2014

Installing Apache Knox

Hortonworks Data Platform

Course Outline. Module 1: Microsoft Azure for AWS Experts Course Overview

Administration 1. DLM Administration. Date of Publish:

At DataWorks 2018, Hortonworks Accelerates Its Shift Toward Public Cloud Deployments by, James Kobielus

FAQs. Business (CIP 2.2) AWS Market Place Troubleshooting and FAQ Guide

Hortonworks Data Platform

Cloudera Director User Guide

/ Cloud Computing. Recitation 5 February 14th, 2017

Container Orchestration on Amazon Web Services. Arun

Installation Guide Revision B. McAfee Cloud Workload Security 5.0.0

Oracle Cloud Using Oracle Big Data Cloud. Release 18.1

Mesosphere and Percona Server for MongoDB. Jeff Sandstrom, Product Manager (Percona) Ravi Yadav, Tech. Partnerships Lead (Mesosphere)

Mesosphere and Percona Server for MongoDB. Peter Schwaller, Senior Director Server Eng. (Percona) Taco Scargo, Senior Solution Engineer (Mesosphere)

Single Sign-On for PCF. User's Guide

Knox Implementation with AD/LDAP

Quick Install for Amazon EMR

/ Cloud Computing. Recitation 5 September 26 th, 2017

ambari administration 2 Administering Ambari Date of Publish:

Index. Scott Klein 2017 S. Klein, IoT Solutions in Microsoft s Azure IoT Suite, DOI /

Known Issues for Oracle Big Data Cloud. Topics: Supported Browsers. Oracle Cloud. Known Issues for Oracle Big Data Cloud Release 18.

ARCHITECTURAL OVERVIEW REVISED 6 NOVEMBER 2018

Module Day Topic. 1 Definition of Cloud Computing and its Basics

Developing Microsoft Azure Solutions (70-532) Syllabus

Cloudera Director User Guide

Installing Apache Atlas

Tech Talk Nutanix Calm. Greg Smith, VP Product Marketing Gil Haberman, Director of Product Marketing

Hadoop. Introduction / Overview

Hortonworks University. Education Catalog 2018 Q1

Building scalable service-based applications Wicked Fast

Altus Data Engineering

Security 3. NiFi Authentication. Date of Publish:

Installing HDF Services on an Existing HDP Cluster

Developing Microsoft Azure Solutions

What is Cloud Computing? What are the Private and Public Clouds? What are IaaS, PaaS, and SaaS? What is the Amazon Web Services (AWS)?

OpenNebula on VMware: Cloud Reference Architecture

Microsoft Azure for AWS Experts

Agile CI/CD with Jenkins and/at ZeroStack. Kiran Bondalapati CTO, Co-Founder & Jenkins Admin ZeroStack, Inc. (

Configuring and Operating a Hybrid Cloud with Microsoft Azure Stack

Developing Microsoft Azure Solutions (70-532) Syllabus

Savanna. Release 0.3.dev4.gcba1bac. OpenStack Foundation

HP Helion CloudSystem 9.0 Administrator Guide

Hortonworks Data Platform

HDI+Talena Resources Deployment Guide. J u n e

Security and Performance advances with Oracle Big Data SQL

Developing Microsoft Azure Solutions (70-532) Syllabus

Installing an HDF cluster

Setting Up Resources in VMware Identity Manager 3.1 (On Premises) Modified JUL 2018 VMware Identity Manager 3.1

Kuberiter White Paper. Kubernetes. Cloud Provider Comparison Chart. Lawrence Manickam Kuberiter Inc

Introduction to data centers

AKAMAI WHITE PAPER. Enterprise Application Access Architecture Overview

Cloudera s Enterprise Data Hub on the AWS Cloud

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

NGF0502 AWS Student Slides

How CloudEndure Works

Security Camp 2016 Cloud Security. August 18, 2016

Installation 1. DLM Installation. Date of Publish:

OnCommand Cloud Manager 3.2 Deploying and Managing ONTAP Cloud Systems

How to Run the Big Data Management Utility Update for 10.1

Ambari User Views: Tech Preview

Docker Universal Control Plane Deploy and Manage On-Premises, Your Dockerized Distributed Applications

Cloudera Director User Guide

Introduction to ArcGIS Server Architecture and Services. Amr Wahba

Swift Web Applications on the AWS Cloud

DevOps and Continuous Delivery USE CASE

August, HPE Propel Microservices & Jumpstart

Real4Test. Real IT Certification Exam Study materials/braindumps

Upgrading Big Data Management to Version Update 2 for Hortonworks HDP

How to Lift-and-Shift a Line of Business Application onto Google Cloud Platform

Installing and Configuring Apache Storm

Course Overview This five-day course will provide participants with the key knowledge required to deploy and configure Microsoft Azure Stack.

vcommander Release GA Release Notes

Configuring Intelligent Streaming 10.2 For Kafka on MapR

Installing SmartSense on HDP

Microsoft Perform Data Engineering on Microsoft Azure HDInsight.

40390: Microsoft Azure for AWS Experts

Introduction to Cloud Computing

Managing and Auditing Organizational Migration to the Cloud TELASA SECURITY

Hortonworks SmartSense

IBM Cloud Orchestrator Version User's Guide IBM

Hortonworks SmartSense

Enabling Cloud Adoption. Addressing the challenges of multi-cloud

Transcription:

2 Introduction to Cloudbreak Date of Publish: 2019-02-06 https://docs.hortonworks.com/

Contents What is Cloudbreak... 3 Primary use cases... 3 Interfaces...3 Core concepts... 4 Architecture... 7 Cloudbreak deployer architecture...7 Cloudbreak application architecture...8

What is Cloudbreak What is Cloudbreak Cloudbreak simplifies the provisioning, management, and monitoring of on-demand HDP and HDF clusters in virtual and cloud environments. It leverages cloud infrastructure to create host instances, and uses Apache Ambari via Ambari blueprints to provision and manage Hortonworks clusters. Cloudbreak allows you to create clusters by using the web UI, CLI, and REST API. Clusters can be launched on public cloud infrastructure platforms Amazon Web Services (AWS) including AWS GovCloud, Microsoft Azure, and Google Cloud Platform (GCP), and on the private cloud infrastructure platform OpenStack. Primary use cases Cloudbreak allows you to create, manage, and monitor your HDP and HDF clusters on your chosen cloud platform: Dynamically deploy, configure, and manage clusters on public and private clouds (AWS, Azure, Google Cloud, OpenStack). Use automated scaling to seamlessly manage elasticity requirements as cluster workloads change. Secure your cluster by enabling Kerberos. Interfaces Cloudbreak includes a web user interface (web UI), command line interface (CLI), and REST APIs which allow you to perform administrative tasks such as registering your cloud provider credentials, registering external sources, and provisioning and managing clusters. The Cloudbreak web UI can be accessed via a browser. The Cloudbreak CLI can be downloaded from Cloudbreak web UI and installed on your computer. 3

Core concepts Core concepts Cloudbreak includes the following core concepts: credentials, clusters, blueprints, cluster extensions, and external sources. Credentials After launching Cloudbreak, you are required to create a Cloudbreak credential for each cloud provider on which you would like to provision clusters. Only after you have completed that step, you can start creating clusters. Cloudbreak credential allows Cloudbreak to authenticate with the cloud provider and obtain authorization to provision cloud provider resources on your behalf. The authentication and authorization process varies depending on the cloud provider, but is typically done via assigning a specific IAM role to Cloudbreak which allows Cloudbreak to perform certain actions within your cloud provider account. 4

Core concepts Clusters Cloudbreak allows you to create two types of clusters: workload clusters (referred to as "workload clusters", "ephemeral clusters", or "clusters") and data lake clusters (referred to as "data lakes"). Workload clusters: Multiple HDP or HDF clusters can be provisioned as needed and then used for storing and processing data. These clusters are typically ephemeral. By default, Cloudbreak includes a few default HDP cluster blueprints that allow you to provision clusters including three node types: master, worker, and compute. A master node runs the components for managing the cluster (including Ambari), storing data, processing tasks, as well as other master components. A worker node runs the components that are used for executing processing tasks and handling storing data in HDFS. A compute node can optionally be used for running data processing tasks. Note: Default HDF cluster blueprints do not follow this pattern. Furthermore, you can upload your own HDP and HDF workload cluster blueprints with custom node types. Data lake clusters: A data lake provides a way for you to centrally apply and enforce authentication, authorization, and audit policies across multiple ephemeral workload clusters. "Attaching" your workload cluster to the data lake instance allows the attached cluster workloads to access data and run in the security context provided by the data lake. Note: The data lake functionality is technical preview. Blueprints Ambari blueprints are a declarative definition of a cluster. A blueprint allows you to specify stack, component layout, and configurations to materialize a cluster instance via Ambari REST API, without having to use the Ambari cluster install wizard. Ambari blueprints are specified in JSON format. After you provide the blueprint to Cloudbreak, the host groups in the JSON are mapped to a set of instances when starting the cluster, and the specified services and components are installed on the corresponding nodes. Cloudbreak includes a few default blueprints and allows you to upload your own blueprints. 5

Core concepts Cluster extensions Cloudbreak allows you to utilize the following cluster extensions: Recipes: Cloudbreak allows you to upload custom scripts, called "recipes". A recipe is a script that runs on all nodes of a selected node group at a specific time. You can use recipes for tasks such as installing additional software or performing advanced cluster configuration. For example, you can use a recipe to put a JAR file on the Hadoop classpath. Management packs: Management packs allow you to deploy a range of services to your Ambari-managed cluster. You can use a management pack to deploy a specific component or service, such as HDP Search, or to deploy an entire platform, such as HDF. External sources Cloudbreak allows you to define external sources that are created independently of a cluster and therefore their lifespan is not limited by the lifespan of any cluster and that can be reused with multiple clusters. The external sources that can be registered in Cloudbreak include: External source Authentication configurations (LDAP/AD) Database configurations Image catalogs Proxy configurations Description Cloudbreak allows you to register an existing LDAP/AD instance as an external source and use it for multiple clusters. Cloudbreak allows you to register existing RDBMS instances as external sources to be used as databases for specific cluster components (such as Ambari, Ranger, Hive, and so on), and then reuse them for multiple clusters. Cloudbreak provides a default image catalog with default prewarmed and base images, and allows you to create custom image catalogs with custom base images. While base images include only OS and tools, prewarmed images also include Ambari and stack. Cloudbreak allows you to register an existing proxy to be used for your clusters. 6

Architecture Workspaces When Cloudbreak instance is configured with an LDAP or Active Directory, the LDAP/AD users can share Cloudbreak resources with other users via workspaces. Any user can create a workspace and invite other users to that workspace, granting them specific access rights. Once a workspace is created, users who are part of it share all resources, such as clusters, blueprints, recipes, and so on, created as part of that workspace. Note: It is not possible to share resources between workspaces; That is, every resource is attached to the workspace in which it was created and it cannot be shared with or moved to another workspace. Architecture Cloudbreak deployer installs Cloudbreak components on a VM. Once these components are deployed, you can use Cloudbreak application or Cloudbreak CLI to create, manage, and monitor clusters. To learn more, refer to the following documentation: Cloudbreak deployer architecture Cloudbreak deployer installs Cloudbreak components on a VM. Cloudbreak deployer includes the following components: Component Cloudbreak application Uluwatu Description Cloudbreak application is built on the foundation of cloud provider APIs and Apache Ambari. This is Cloudbreak web UI, which can be used to create, manage, and monitor clusters. 7

Architecture Component Cloudbreak CLI Identity Sultans Periscope Description This is Cloudbreak s command line tool, which can be used to create, manage, and monitor clusters. This is Cloudbreak s OAuth identity server implementation, which utilizes UAA. This is Cloudbreak s user management system. This is Cloudbreak s autoscaling application, which is responsible for automatically increasing or decreasing the capacity of the cluster when your pre-defined conditions are met. Cloudbreak application architecture Cloudbreak application is a web application which simplifies cluster provisioning in the cloud. Based on your input, Cloudbreak application provisions all required cloud infrastructure and then provisions a cluster on your behalf within your cloud provider account. Cloudbreak application is built on the foundation of cloud provider APIs and Apache Ambari: Cloudbreak uses the Cloudbreak credential to authenticate with your cloud provider account and provision cloud resources required for the clusters. Cloudbreak uses cloud provider APIs to communicate with the cloud providers. Cloudbreak uses Apache Ambari and Ambari blueprints to provision, manage, and monitor clusters. Ambari blueprints are a declarative definition of a cluster. With a blueprint, you can specify stack, component layout, and configurations to materialize a cluster instance via Ambari REST API, without having to use the Ambari cluster install wizard. 8