IMPLEMENTING HTTPFS & KNOX WITH ISILON ONEFS TO ENHANCE HDFS ACCESS SECURITY

Size: px
Start display at page:

Download "IMPLEMENTING HTTPFS & KNOX WITH ISILON ONEFS TO ENHANCE HDFS ACCESS SECURITY"

Transcription

1 IMPLEMENTING HTTPFS & KNOX WITH ISILON ONEFS TO ENHANCE HDFS ACCESS SECURITY Boni Bruno, CISSP, CISM, CGEIT Principal Solutions Architect DELL EMC ABSTRACT This paper describes implementing HTTPFS and Knox together with Isilon OneFS to enhance HDFS access security. This integrated solution has been tested and certified with Hortonworks on HDP v2.4 and Isilon OneFS v

2 CONTENTS Introduction... 3 WebHDFS REST API... 3 WebHDFS Port Assignment in Isilon OneFS... 5 WebHDFS Examples with ISILON... 5 WebHDFS Security Concerns... 8 HTTPFS... 9 Installing HTTPFS... 9 Configuring HTTPFS Configuring HTTPFS for Kerberos Running and Stopping HTTPFS Configuring HTTPFS Auto-Start Testing HTTPFS Knox Installing Knox Configuring Knox using Ambari Configuring Knox for LDAP Configuring Knox for Kerberos Testing Knox and Isilon Impersonation Defense Final Comments Appendix Additional Testing Results... 38

3 INTRODUCTION Hadoop provides a Java native API to support file system operations such as create, rename or delete files and directories, open, read or write files, set permissions, etc. This is great for applications running within the Hadoop cluster, but there may be use cases where an external application needs to make such file system operations on files stored on HDFS as well. Hortonworks developed the WebHDFS REST API to support these requirements based on standard REST functionalities. WebHDFS REST APIs support a complete File System / File Context interface for HDFS. WEBHDFS REST API WEBHDFS IS BASED ON HTTP OPERATIONS LIKE GET, PUT, POST AND DELETE. WEBHDFS OPERATIONS LIKE OPEN, GETFILESTATUS, LISTSTATUS ARE USING HTTP GET, OTHER OPERATIONS LIKE CREATE, MKDIRS, RENAME, SETPERMISSIONS ARE RELYING ON HTTP PUT. APPEND OPERATIONS ARE BASED ON HTTP POST, WHILE DELETE IS USING HTTP DELETE. AUTHENTICATION CAN BE BASED ON USER NAME, QUERY PARAMETER (AS PART OF THE HTTP QUERY STRING) OR IF SECURITY IS ENABLED, THROUGH KERBEROS. Web HDFS is enabled in a Hadoop cluster by defining the following property in hdfs-site.xml: <property> <name>dfs.webhdfs.enabled</name> <value>true</value> </property> If using Ambari, enable WebHDFS under the General Settings of HDFS as shown below:

4 When using Isilon as a centralized HDFS storage repository for a given Hadoop Cluster, all namenode and datanode functions must be configured to run on Isilon for the entire Hadoop cluster. By design, WebHDFS needs access to all nodes in the cluster. Before the WebHDFS interface on Isilon can be used by the Hadoop Cluster, you must enable WebHDFS in the Protocol Settings for HDFS on the designated Access Zone in Isilon - this is easily done in the OneFS GUI. In the example below, hdp24 is the HDFS Access Zone for the Hadoop Cluster. Note the check mark next to ENABLE WebHDFS access. It is not sufficient to just enable WebHDFS in Ambari. Isilon must also be configured with WebHDFS enabled so end to end WebHDFS communication can work in the Hadoop cluster. If multiple Access Zones are defined on Isilon, make sure to enable WebHDFS as needed on each access zone.

5 WEBHDFS PORT ASSIGNMENT IN ISILON ONEFS All references to Hadoop host hdp24 in this document refer to a defined SmartConnect HDFS Access Zone on Isilon. TCP Port 8082 is the port OneFS uses for WebHDFS. It is important that the hdfs-site.xml file in the Hadoop Cluster reflect the correct port designation for HTTP access to Isilon. See Ambari screen shot below for reference. WEBHDFS EXAMPLES WITH ISILON Assuming the Hadoop cluster is up and running with Isilon and WebHDFS has been properly enabled for the Hadoop cluster, we are ready to test WebHDFS. CURL is a great command line tool for transferring data using various protocols, including HTTP/HTTPS. The examples below use curl to invoke the WebHDFS REST API available in Isilon OneFS to conduction various file system operations. Again, all references to hdp24 used in the curl commands below refer to the SmartConnect HDFS Access Zone on Isilon and not some edge node in the cluster. GETTING FILE STATUS EXAMPLE The screen shot above shows curl being used to connect to Isilon s WebHDFS interface on port 8082, the GETFILESTATUS operation is used as user hduser1 to retrieve info on the projects.txt file. Note: The projects.txt file is a test file I created. It is not part of the Hortonworks software.

6 A web browser may also be used to get projects.txt file status from Isilon WebHDFS as shown below: This is similar to executing hdfs dfs ls /user/hduser1/projects.txt from a Hadoop client node n107 as shown below: This quick example shows the flexibility of using WebHDFS. It provides a simple way to execute Hadoop file system operations by an external client that does not necessarily run on the Hadoop cluster itself. Let s look at another example. CREATING A DIRECTORY EXAMPLE Here the MKDIRS operation on a different client node n105 is used with PUT to create the directory /tmp/hduser as user hduser1 on Isilon. We can tell by the true Boolean result the operation was successful. We can also check the result by using hdfs to see the directory on Isilon as shown below:

7 OPEN A FILE EXAMPLE In the example above, the OPEN operation is used with curl to display the text string Knox HTTPFS Isilon Project within the /tmp/hduser1/project.txt file. As shown before, a web browser can be used to access the file as well. Here a browser is configured to automatically open text files in notepad, so accessing the WebHDFS API on Isilon as shown below will open the contents of /tmp/hduser1/project.txt in notepad directly. To validate the contents from within the cluster, we can use hdfs as show below:

8 I m only scratching the surface with the examples above; there are various operations you can execute with WebHDFS. You can easily use WebHDFS to append data to files, rename files or directories, create new files, etc. See the Appendix for many more examples. It should be apparent that WebHDFS provides a simple, standard way to execute Hadoop file system operations with external clients that do not necessarily run within the Hadoop cluster itself. WEBHDFS SECURITY CONCERNS SOMETHING WORTH POINTING OUT WITH THE ABOVE EXAMPLES AND WITH WEBHDFS IN GENERAL CLIENTS ARE DIRECTLY ACCESSING THE NAMENODES AND DATANODES VIA PREDEFINED PORTS. THIS CAN BE SEEN AS A SECURITY ISSUE FOR MANY ORGANIZATIONS WANTING TO ENABLE EXTERNAL WEBHDFS ACCESS TO THEIR HADOOP INFRASTRUCTURE. Many organizations do not want their Hadoop infrastructure accessed directly from external clients. As seen thus far, external clients can use WebHDFS to directly access the actual ports namenodes and datanodes are listening on in the Hadoop Cluster and leverage the WebHDFS REST API to conduct various file system operations. Although firewalls can filter access from external clients, the ports are still being directly access. As a result, firewalls do not prohibit the execution of various WebHDFS operations. The solution to this issue, in many cases, is to enable Kerberos in the Hadoop cluster and deploy Secure REST API Gateways that enforce strong authentication and access control to WebHDFS. The remainder of this document focuses on using HTTPFS and Knox in conjunction with Isilon OneFS to provide a secure WebHDFS deployment with Hadoop. A diagram of the secure architecture is shown below for reference.

9 HTTPFS The introduction section of this document provides an overview of WebHDFS and demonstrates how the WebHDFS REST APIs support a complete File System / File Context interface for HDFS. WebHDFS is efficient as it streams data from each datanode and can support external clients like curl or web browsers to extend data access beyond the Hadoop cluster. Since WebHDFS needs access to all nodes in the cluster by design, WebHDFS inherently establishes a wider foot print for HDFS access in a Hadoop cluster since clients can access HDFS over HTTP/HTTPS. To help minimize the size of the foot print to clients, a gateway solution is needed that provides a similar File System / File Context interface for HDFS, and this is where HTTPFS comes in to play. HTTPFS is a service that provides a REST HTTP gateway supporting all HDFS File System operations (read and write). HTTPFS can be used to provide a gateway interface, i.e. choke point, to Isilon and limit broad HDFS access from external clients to the Hadoop cluster. HTTPFS can also be integrated with Knox to improve service level authorization, LDAP & AD integration, and overall perimeter security. See the Knox section of this document for more details. The remainder of this section covers the installation and configuration of HTTPFS with Isilon. INSTALLING HTTPFS HTTPFS can be installed on Ambari server or a worker node, for production deployments, deploying on a dedicated worker node is a best practice. To install HTTPFS: yum install hadoop-httpfs (Note: existing HWX repos are hadoop-httpfs aware) Note: The HTTPFS service is a tomcat application that relies on having the Hadoop libraries and configuration available, so make sure to install HTTPFS on an edge node that is being managed by Ambari. After you install HTTPFS, the directories below will be created on the HTTPFS server: /usr/hdp/2.x.x.x-x/hadoop-httpfs /etc/hadoop-httpfs/conf /etc/hadoop-httpfs/tomcat-deployment

10 CONFIGURING HTTPFS If you change directories to /usr/hdp on your HTTPFS server and list the files there, you will see a directory with the version number of your existing HDP release. Make note of it so you can set the current version for httpfs. Set the version for current with the following command: hdp-select set hadoop-httpfs 2.x.x.x-x (replace the x with your HDP release) The installation of httpfs above deploys scripts which have some hardcoded values that need to be changed. Adjust the /usr/hdp/current/hadoop-httpfs/sbin/httpfs.sh script: #!/bin/bash # Autodetect JAVA_HOME if not defined if [ -e /usr/libexec/bigtop-detect-javahome ]; then. /usr/libexec/bigtop-detect-javahome elif [ -e /usr/lib/bigtop-utils/bigtop-detect-javahome ]; then. /usr/lib/bigtop-utils/bigtop-detect-javahome fi ### Added to assist with locating the right configuration directory export HTTPFS_CONFIG=/etc/hadoop-httpfs/conf ### Remove the original HARD CODED Version reference. Next, you need to create the following symbolic links: cd /usr/hdp/current/hadoop-httpfs ln -s /etc/hadoop-httpfs/tomcat-deployment/conf conf ln -s../hadoop/libexec libexec

11 Like all the other Hadoop components, httpfs follows the use of *-env.sh files to control the startup environment. Above, in the httpfs.sh script we set the location of the configuration directory, this configuration directory is used to find and load the httpfs-env.sh file. The httpfs-env.sh file needs to be modified as shown below: # Add exports to control and set the Catalina directories for starting and finding the httpfs application export CATALINA_BASE=/usr/hdp/current/hadoop-httpfs export HTTPFS_CATALINA_HOME=/etc/hadoop-httpfs/tomcat-deployment # Set a log directory that matches your standards export HTTPFS_LOG=/var/log/hadoop/httpfs # Set a tmp directory for httpfs to store interim files export HTTPFS_TEMP=/tmp/httpfs The default port for httpfs is TCP If you need to change the port for httpfs, add the following export to the above httpfs-env.sh file on the HTTPFS server: export HTTPFS_HTTP_PORT=<new_port> In the Ambari web interface, add httpfs as a proxy user in core-site.xml in the HDFS > Configs > Advanced > Custom core site section: Note: If the properties that are referenced below do not already exist, do the following steps: 1. Click the Add Property link in the Custom core site area to open the Add Property window. 2. Add each value in the <name> part in the Key field. 3. Add each value in the <value> part in the Value field. 4. Click Add. Then click Save.

12 <property> <name>hadoop.proxyuser.httpfs.groups</name> <value>*</value> </property> <property> <name>hadoop.proxyuser.httpfs.hosts</name> <value>*</value> </property> Make sure to restart HDFS and related components after making the above changes to core-site.xml. At this point HTTPFS is configured to work with a non-kerberos Hadoop cluster. If your cluster is not secured with Kerberos, you can skip the following section CONFIGURING HTTPFS FOR KERBEROS and proceed to RUNNING AND STOPPING HTTPFS and TESTING HTTPFS.

13 CONFIGURING HTTPFS FOR KERBEROS Ambari does not automate the configuration of HTTPFS to support Kerberos. If your Hadoop cluster was secured with Kerberos using Ambari, you will need to create some needed keytabs and modify the httpfs-site.xml before HTTPFS will work in a secure Kerberos Hadoop cluster. The following assumptions are made for this section on configuring HTTPFS for Kerberos: 1. HTTPFS has been installed, configured, and verified to be working prior to enabling Kerberos. 2. Kerberos was enabled using Ambari and an MIT KDC and Isilon is configured and verified for Kerberos. Both httpfs and HTTP service principals must be created for HTTPFS if they do not already exist. Create the httpfs and HTTP (see note below) principals: kadmin: addprinc -randkey httpfs/fully.qualified.domain.name@example-realm.com kadmin: addprinc -randkey HTTP/fully.qualified.domain.name@EXAMPLE-REALM.COM Note: HTTP principal and keytab may already exist as this is typically needed for other Hadoop services in a secure Kerberos Hadoop cluster deployment. HTTP must be in CAPITAL LETTERS. Create the keytab files for both httpfs and HTTP (see note above) principals: kadmin -q "ktadd -k /etc/security/keytabs/httpfs.service.keytab httpfs/fully.qualified.domain.name@example-realm.com" kadmin -q "ktadd -k /etc/security/keytabs/spnego.service.keytab HTTP/fully.qualified.domain.name@EXAMPLE-REALM.COM" Note: The spnego keytab above only needs to be created if it does not already exist on the node running HTTPFS. Merge the two keytab files into a single keytab file:

14 ktutil: rkt /etc/security/keytabs/httpfs.service.keytab ktutil: rkt /etc/security/keytabs/spnego.service.keytab ktutil: wkt /etc/security/ketyabs/httpfs-http.service.keytab ktutil: quit The above will create a file named httpfs-http.service.keytab in /etc/security/keytabs Test that the merged keytab file works: klist -kt /etc/security/keytabs/httpfs-http.service.keytab The above command should list both hdfs and HTTP principals for the httpfs-http.service.keytab. Below is an example output from a test cluster: Change the ownership and permissions of the /etc/security/keytabs/httpfs-http.service.keytab file: chown httpfs:hadoop /etc/security/keytabs/httpfs-http.service.keytab chmod 400 /etc/security/keytabs/httpfs-http.service.keytab Edit the HTTPFS server httpfs-site.xml configuration file in the HTTPFS configuration directory by setting the following properties:

15 httpfs.authentication.type: kerberos httpfs.hadoop.authentication.type: kerberos httpfs.authentication.kerberos.principal: YOUR-REALM.COM> httpfs.authentication.kerberos.keytab: /etc/hadoop-httpfs/conf/httpfs-http.keytab httpfs.hadoop.authentication.kerberos.principal: YOUR-REALM.COM> httpfs.hadoop.authentication.kerberos.keytab: /etc/security/keytabs/httpfs-http.keytab httpfs.authentication.kerberos.name.rules: Use the value configured for 'hadoop.security.auth_to_local' in Ambari's HDFS Configs under "Advanced Core-Site". An example httpfs-site.xml is listed below, with the relevant Kerberos information highlighted in red: <configuration> <!-- HTTPFS proxy user setting --> <property> <name>httpfs.proxyuser.knox.hosts</name> <value>*</value> </property> <property> <name>httpfs.proxyuser.knox.groups</name> <value>*</value> </property> <!-- HUE proxy user setting --> <property> <name>httpfs.proxyuser.hue.hosts</name> <value>*</value>

16 </property> <property> <name>httpfs.proxyuser.hue.groups</name> <value>*</value> </property> <property> <name>httpfs.hadoop.config.dir</name> <value>/etc/hadoop/conf</value> </property> <property> <name>httpfs.authentication.type</name> <value>kerberos</value> </property> <property> <name>httpfs.hadoop.authentication.type</name> <value>kerberos</value> </property> <property> <name>kerberos.realm</name> <value>solarch.lab.emc.com</value> </property>

17 <property> <name>httpfs.authentication.kerberos.principal</name> </property> <property> <name>httpfs.authentication.kerberos.keytab</name> <value>/etc/security/keytabs/httpfs-http.service.keytab</value> </property> <property> <name>httpfs.hadoop.authentication.kerberos.principal</name> </property> <property> <name>httpfs.hadoop.authentication.kerberos.keytab</name> <value>/etc/security/keytabs/httpfs-http.service.keytab</value> </property> <property> <name>httpfs.authentication.kerberos.name.rules</name> <value>

18 DEFAULT </value> </property> </configuration> This concludes the configuration work needed for HTTPFS to work in a secure Kerberos Hadoop cluster. Follow the instructions in the next sections to start and test HTTPFS with Isilon.

19 RUNNING AND STOPPING HTTPFS Executing httpfs is simple. To start: cd /usr/hdp/current/hadoop-httpfs/sbin./httpfs.sh start To stop:./httpfs.sh stop CONFIGURING HTTPFS AUTO-START As the root user, create the following hadoop-httpfs script in /etc/init.d: #!/bin/bash hdp-select set hadoop-httpfs 2.x.x.x.x-x # See how we were called. case "$1" in start) /usr/hdp/current/hadoop-httpfs/sbin/httpfs.sh start ;; stop) /usr/hdp/current/hadoop-httpfs/sbin/httpfs.sh stop ;; *) echo $"Usage: $prog {start stop restart}" esac

20 As root user: chmod 755 /etc/init.d/hadoop-httpfs chkconfig --add hadoop-httpfs # Start Service service hadoop-httpfs start # Stop Service service hadoop-httpfs stop This method will run the service as the httpfs user. Ensure that the httpfs user has permissions to write to the log directory /var/log/hadoop/httpfs. The correct permission settings are shown below: Note: the httpfs user also needs to be created on Isilon. The httpfs user is a system account that gets created during installation of httpfs. As with all other Hadoop server accounts, Isilon needs to have all service accounts defined as a LOCAL PROVIDER in the appropriate HDFS Access Zone (e.g. hdp24) as shown below.

21 Create the httpfs user in the LOCAL HDFS Access Zone for your cluster in Isilon OneFS. Assign the httpfs user to the hadoop primary group. Leave the httpfs account Disabled as shown above and below. The UID on Isilon does not need to match the UID on the httpfs server.

22 TESTING HTTPFS As seen in the introduction section of this document, the curl command is an excellent tool for testing WebHDFS; the same is true for testing HTTPFS. The default port for httpfs is TCP PORT The tests below show how HTTPFS and Isilon OneFS can be used together in a Hadoop cluster. The requests made on port on the HTTPFS gateway are passed to Isilon. The HTTPFS gateway is configured for Kerberos as is the Isilon HDFS Access Zone. The Kerberos configuration is optional, but recommended for production Hadoop deployments to improve cluster security. The testing below is with Kerberos enabled. So make sure you have obtained and cached an appropriate Kerberos ticket-granting ticket before running the commands. Use klist to verify you have a ticket cached as shown below: GETTING A USER S HOME DIRECTORY EXAMPLE The screen shot above shows curl being used to connect to the HTTPFS gateway on port14000, the GETHOMEDIRECTORY operation is used on user hduser1 to retrieve the home directory info. HTTP enables GSS-Negotiate authentication. It is primarily meant as a support for Kerberos5 authentication but may be also used along with another authentication method. GSS-Negotiate is specified with the -negotiate option with curl and the w defines what to display on stdout after a completed and successful operation.

23 LIST DIRECTORY EXAMPLE The screen shot above shows curl being used to connect to the HTTPFS gateway on port14000, the LISTSTATUS operation is used as user hduser1 to do a directory listing on /tmp/hduser1. CREAT DIRECTORY EXAMPLE The screen shot above shows curl being used to connect to the HTTPFS gateway on port14000, the MKDIRS operation is used as user hduser1 to create the directory /tmp/hduser1/test. The Boolean result of true means the command executed successfully. We can verify the creation of the directory with the hdfs command as show below: This concludes the HTTPFS installation, configuration, and testing section of this document. The next section covers how to integrate Knox with HTTPFS and Isilon.

24 KNOX Knox enables the integration of enterprise identity management solutions and numerous perimeter security features for REST/HTTP access to Hadoop and provides perimeter security for Hadoop services. Knox currently supports YARN, WebHCAT, Oozie, HBase, Hive, and WebHDFS Hadoop services. The focus of this paper is on the WebHDFS Hadoop service only. Just like HTTPFS, Knox can be installed on Kerberized and Non-Kerberized Hadoop clusters. Knox by default uses WebHDFS to perform any HDFS operation, but it can also leverage HTTPFS for the same HDFS operations. Knox with HTTPFS provides a defense in depth strategy around REST/HTTP access to Hadoop and Isilon OneFS. This section covers the installation and configuration of Knox and LDAP services to work with HTTPFS in a Kerberized cluster to provide secure REST/HTTP communications to Hadoop and Isilon OneFS. INSTALLING KNOX Knox is included with Hortonworks Data Platform by default. If you unselected the Knox service during installation of HDP, just click the Actions button in Ambari and select the Knox service as shown below and click install. CONFIGURING KNOX USING AMBARI Knox can be managed through Ambari. Since HTTPFS runs on port 14000, a topology change to Knox for the WebHDFS role is needed. Change the topology within the Advance topology section in Ambari/Knox, an example topology configuration for the WebHDFS role is shown below:

25 The WebHDFS role is listed as a service in the topology configuration: <service> <role>webhdfs</role> <url> </service> The HTTPFS_HOST should be replaced with the fully qualified name of the HTTPFS server. Port is the default port for HTTPFS. If you made a change to the HTTPFS port assignment make sure to reflect the port change in the Knox topology configuration as well. Everything else in the topology configuration can be left alone unless you made other port changes to other services. In the Ambari web interface, check that knox is configured as a proxy user in core-site.xml in the HDFS > Configs > Advanced > Custom core site section and that the fully qualified domain name of the Knox host is set. Note: If the properties that are referenced below do not already exist, do the following steps: 1. Click the Add Property link in the Custom core site area to open the Add Property window. 2. Add each value in the <name> part in the Key field. 3. Add each value in the <value> part in the Value field. 4. Click Add. Then click Save.

26 <property> <name>hadoop.proxyuser.knox.host</name> <value>n105.solarch.lab.emc.com</value> </property> <property> <name>hadoop.proxyuser.knox.groups</name> <value>users</value> </property> Make sure to restart HDFS and related components after making the above changes to core-site.xml. CONFIGURING KNOX FOR LDAP Knox can easily integrate with LDAP - just add an LDAP provider and associated parameters to the topology configuration and you are done. An example LDAP provider (within the topology file) is shown below: <provider> <role>authentication</role> <name>shiroprovider</name> <enabled>true</enabled> <param> <name>sessiontimeout</name> <value>30</value> </param> <param> <name>main.ldaprealm</name> <value>org.apache.hadoop.gateway.shirorealm.knoxldaprealm</value> </param> <param> <name>main.ldaprealm.userdntemplate</name>

27 <value>uid={0},ou=people,dc=hadoop,dc=apache,dc=org</value> </param> <param> <name>main.ldaprealm.contextfactory.url</name> <value>ldap://localhost:33389</value> </param> <param> <name>main.ldaprealm.contextfactory.authenticationmechanism</name> <value>simple</value> </param> <param> <name>urls./**</name> <value>authcbasic</value> </param> </provider> The LDAP provider directs Knox to use a directory service for authentication. In the example above, a local LDAP Provider (port 33389) is being used for basic authentication for all urls. Make sure you use a supported LDAP service compatible with Hortonworks and Isilon and to modify the Knox topology configuration to match your deployed LDAP configuration if LDAP will be used with Knox. Supported LDAP servers: OpenLDAP Active Directory w/ RFC2307 schema extension Apple OpenDirectory (OD) Centrify Oracle Directory Sever ApacheDS Red Hat Directory Server (RHDS) Radiantlogic VDS Novell Directory Server (NDS)

28 CONFIGURING KNOX FOR KERBEROS If the Hadoop cluster is secure with Kerberos, you need to make sure Knox is configured for Kerberos as well to avoid authentication errors with the HTTPFS gateway and backend Isilon cluster. The Kerberos configuration for Knox is done under Advance gateway-site in Ambari. An example configuration is shown below: The Advanced gateway-site configuration allows you to specify the Knox gateway port (e.g. 8444), the location of the krb5.conf (Kerberos configuration file), and set the gateway to use Kerberos (set to true). The Advance knox-env in Ambari allows you to set the Knox user and group accounts, Knox keytab path, and Knox Principal Name. An example configuration is shown below: Note: the knox user also needs to be created on Isilon. The knox user is a system account that gets created during installation of knox. As with all other Hadoop server accounts, Isilon needs to have all service accounts defined as a LOCAL PROVIDER in the appropriate HDFS Access Zone (e.g. hdp24) as shown below.

29 Create the knox user in the LOCAL HDFS Access Zone for your cluster in Isilon OneFS. Assign the knox user to the hadoop primary group. Leave the knox account Disabled as shown above and below. The UID on Isilon does not need to match the UID on the knox server.

30 TESTING KNOX AND ISILON IMPERSONATION DEFENSE Now that Knox and HTTPFS have been installed and configured, we can begin end-to-end testing with Isilon in a secure Kerberos Hadoop cluster deployment using either curl or a web browser. GETTING A USER S HOME DIRECTORY EXAMPLE The screen shot above shows curl being used to connect to the Knox gateway on port 8444 with LDAP user ldapuser1, the GETHOMEDIRECTORY operation is used to retrieve the home directory info for the LDAP user. The network connection to the Knox gateway is secured with TLS. Let s see what happens when we use the same REST HTTP operation over a web browser that connects to the Knox gateway: First, the Knox gateway will prompt for user authentication, after entering the correct LDAP credentials, we can see the result of the REST HTTP GETHOMEDIRECTORY operation in the web browser as shown below:

31 Note that the network connection to the Knox gateway is secured with TLS as shown below: I used self-signed certificates for this lab deployment, so there is a certificate error shown, but the network connection is securely encrypted with TLS and a strong AES cipher. OPENING A FILE EXAMPLE Unlike the GETHOMEDIRECTORY operation shown in the previous test example, the OPEN operation actually accesses data - we want to employ more security checks when data is being access in cases like this.

32 The screen shot above shows curl being used to connect to the Knox gateway on port 8444 as LDAP user ldapuser1, the OPEN operation then tries to open the contents of the project.txt file in /tmp/hduser1, but a Server Error is encountered. Although Isilon is aware of ldapuser1, Isilon provides an added layer of security to check for impersonation attacks. In this case, the HTTPFS gateway (which runs as the httpfs user) is acting as a proxy for user ldapuser1 REST HTTP request between Knox and Isilon. When Isilon receives the OPEN request from httpfs on behalf of ldapuser1, Isilon checks its Proxy User settings to see if httpfs is authorized to impersonate as ldapuser1 or the group ldapuser1 is in, i.e. the hadoop group. Assuming it is within policy for httpfs to impersonate anyone in the hadoop group, we can update the Proxy User settings on Isilon so httpfs is authorized to process requests from either the ldapuser1 user specifically or anyone in the hadoop group. The example below depicts a proxy configuration for the hadoop group: With the proxy user setting in place, we can successfully run the previous test example to open a file:

33 As show above, with the correct Isilon Proxy User Policy in place on Isilon, the Open operation is now allowed. Note: If the /tmp/hduser1 directory on Isilon did not have global read permissions set, this operation would fail as shown below: Changing the permissions on the /tmp/hduser1 directory on Isilon caused a permission denied error for the same previous test operation. This is a testament to the embedded Isilon OneFS security features and a benefit of using a centralized HDFS storage solution like Isilon. CREAT DIRECTORY EXAMPLE The screen shot above shows curl being used to connect to the Knox gateway on port 8444, the MKDIRS operation is used as user ldapuser1 to create the directory /tmp/ldaptest. The Boolean result of true means the command executed successfully.

34 We can verify the creation of the directory with the hdfs command as show below: This concludes the Knox installation, configuration, and testing section of this document. Please see the Appendix for additional Knox/HTTPFS/Isilon test examples.

35 FINAL COMMENTS This solution has been tested and certified by both DELL EMC and Hortonworks with success. One thing that was noticed during testing of the integrated solution is that httpfs wants the header content-type: octet stipulated on data upload requests. The content-type is support by both WebHDFS & HTTPFS, but HTTPFS will throw a 400 Bad Request Error. For example, let say you create a test data_file on the cluster with the CREATE operation, you will need to use the H flag with curl to specify the Content-Type accordingly, see example below: With the Content-Type specified, the data upload successfully completes with no errors. This is an HTTPFS requirement and has nothing to do with either Knox or Isilon OneFS. We can use hdfs command to see the content of the created data_file as shown below:

36 Reading the file via curl does not require anything special as shown below: The port for Knox was changed to 8444 instead of the default Be aware when setting up HTTPS for the Ambari web interface, the default port is also To avoid port conflicts, I recommend you carefully assign a unique port to your Knox gateway; port 8444 is a safe bet.

37 APPENDIX

38 ADDITIONAL TESTING RESULTS Below are additional testing examples for reference. RENAMING A FILE EXAMPLE The above curl command connects to the Knox gateway on port 8444 as LDAP user ldapuser1 to execute the RENAME operation to rename data_file to data_file_new, the Boolean result of true means the command executed successfully. We can verify further my listing the contents of the /tmp/ldaptest directory: SETTING FILE REPLICATION EXAMPLE The above curl command connects to the Knox gateway on port 8444 as LDAP user ldapuser1 to execute the SETREPLICATION operation to set replication to 1 for data_file_new, the Boolean result of true means the command executed successfully.

39 Note: Isilon will always respond with true for these kinds of requests, but the reality is that Isilon OneFS file system is much more efficient than HDFS, Isilon uses erasure encoding instead of replication to maintain high availability. SETTING FILE PERMISSIONS EXAMPLE The above curl command connects to the Knox gateway on port 8444 as LDAP user ldapuser1 to execute the SETPERMISSION operation to 777 for data_file_new, the HTTP/ OK result of means the command executed successfully. The hdfs command shows that the permissions for this data file were changed on Isilon accordingly. APPENDING DATA TO A FILE EXAMPLE

40 The above curl command connects to the Knox gateway on port 8444 as LDAP user ldapuser1 to execute the APPEND operation to add ApendInfo to data_file_new, the HTTP/ OK result of means the command executed successfully. The hdfs command shows the data was appended successfully on Isilon. RECURSIVE DELETE EXAMPLE The above curl command connects to the Knox gateway on port 8444 as LDAP user ldapuser1 to execute the DELETE operation to recursively delete from /tmp/ldaptest on, the HTTP/ OK result of means the command executed successfully. The hdfs command shows the directory and its content was successfully removed from Isilon.

ISILON ONEFS WITH HADOOP KERBEROS AND IDENTITY MANAGEMENT APPROACHES. Technical Solution Guide

ISILON ONEFS WITH HADOOP KERBEROS AND IDENTITY MANAGEMENT APPROACHES. Technical Solution Guide ISILON ONEFS WITH HADOOP KERBEROS AND IDENTITY MANAGEMENT APPROACHES Technical Solution Guide Hadoop and OneFS cluster configurations for secure access and file permissions management ABSTRACT This technical

More information

Configuring Apache Knox SSO

Configuring Apache Knox SSO 3 Configuring Apache Knox SSO Date of Publish: 2018-07-15 http://docs.hortonworks.com Contents Configuring Knox SSO... 3 Configuring an Identity Provider (IdP)... 4 Configuring an LDAP/AD Identity Provider

More information

Knox Implementation with AD/LDAP

Knox Implementation with AD/LDAP Knox Implementation with AD/LDAP Theory part Introduction REST API and Application Gateway for the Apache Hadoop Ecosystem: The Apache Knox Gateway is an Application Gateway for interacting with the REST

More information

TIBCO Spotfire Connecting to a Kerberized Data Source

TIBCO Spotfire Connecting to a Kerberized Data Source TIBCO Spotfire Connecting to a Kerberized Data Source Introduction Use Cases for Kerberized Data Sources in TIBCO Spotfire Connecting to a Kerberized Data Source from a TIBCO Spotfire Client Connecting

More information

Installing SmartSense on HDP

Installing SmartSense on HDP 1 Installing SmartSense on HDP Date of Publish: 2018-07-12 http://docs.hortonworks.com Contents SmartSense installation... 3 SmartSense system requirements... 3 Operating system, JDK, and browser requirements...3

More information

Ambari User Views: Tech Preview

Ambari User Views: Tech Preview Ambari User Views: Tech Preview Welcome to Hortonworks Ambari User Views Technical Preview. This Technical Preview provides early access to upcoming features, letting you test and review during the development

More information

Configuring Apache Knox SSO

Configuring Apache Knox SSO 3 Configuring Apache Knox SSO Date of Publish: 2018-07-15 http://docs.hortonworks.com Contents Setting Up Knox SSO...3 Configuring an Identity Provider (IdP)... 3 Configuring an LDAP/AD Identity Provider

More information

HDP Security Audit 3. Managing Auditing. Date of Publish:

HDP Security Audit 3. Managing Auditing. Date of Publish: 3 Managing Auditing Date of Publish: 2018-07-15 http://docs.hortonworks.com Contents Audit Overview... 3 Manually Enabling Audit Settings in Ambari Clusters...3 Manually Update Ambari Solr Audit Settings...3

More information

Installation 1. DLM Installation. Date of Publish:

Installation 1. DLM Installation. Date of Publish: 1 DLM Installation Date of Publish: 2018-05-18 http://docs.hortonworks.com Contents Installation overview...3 Setting Up the Local Repository for Your DLM Installation... 3 Set up a local repository for

More information

HDP Security Overview

HDP Security Overview 3 HDP Security Overview Date of Publish: 2018-07-15 http://docs.hortonworks.com Contents HDP Security Overview...3 Understanding Data Lake Security... 3 What's New in This Release: Knox... 5 What's New

More information

HDP Security Overview

HDP Security Overview 3 HDP Security Overview Date of Publish: 2018-07-15 http://docs.hortonworks.com Contents HDP Security Overview...3 Understanding Data Lake Security... 3 What's New in This Release: Knox... 5 What's New

More information

Boni Bruno, Chief Solutions Architect, EMC BLUETALON AUDITING AND AUTHORIZATION WITH HDFS ON ISILON ONEFS V8.0

Boni Bruno, Chief Solutions Architect, EMC BLUETALON AUDITING AND AUTHORIZATION WITH HDFS ON ISILON ONEFS V8.0 Boni Bruno, Chief Solutions Architect, EMC BLUETALON AUDITING AND AUTHORIZATION WITH HDFS ON ISILON ONEFS V8.0 1 Secure, Fast, Flexible Hadoop Data Security Solution for Enterprises Analyze data at any

More information

Hortonworks Technical Preview for Apache Falcon

Hortonworks Technical Preview for Apache Falcon Architecting the Future of Big Data Hortonworks Technical Preview for Apache Falcon Released: 11/20/2013 Architecting the Future of Big Data 2013 Hortonworks Inc. All Rights Reserved. Welcome to Hortonworks

More information

Hadoop Security. Building a fence around your Hadoop cluster. Lars Francke June 12, Berlin Buzzwords 2017

Hadoop Security. Building a fence around your Hadoop cluster. Lars Francke June 12, Berlin Buzzwords 2017 Hadoop Security Building a fence around your Hadoop cluster Lars Francke June 12, 2017 Berlin Buzzwords 2017 Introduction About me - Lars Francke Partner & Co-Founder at OpenCore Before that: EMEA Hadoop

More information

Installing Apache Zeppelin

Installing Apache Zeppelin 3 Installing Date of Publish: 2018-04-01 http://docs.hortonworks.com Contents Install Using Ambari...3 Enabling HDFS and Configuration Storage for Zeppelin Notebooks in HDP-2.6.3+...4 Overview... 4 Enable

More information

KNIME Extension for Apache Spark Installation Guide. KNIME AG, Zurich, Switzerland Version 3.7 (last updated on )

KNIME Extension for Apache Spark Installation Guide. KNIME AG, Zurich, Switzerland Version 3.7 (last updated on ) KNIME Extension for Apache Spark Installation Guide KNIME AG, Zurich, Switzerland Version 3.7 (last updated on 2018-12-10) Table of Contents Introduction.....................................................................

More information

Installing Apache Atlas

Installing Apache Atlas 3 Installing Apache Atlas Date of Publish: 2018-04-01 http://docs.hortonworks.com Contents Apache Atlas prerequisites... 3 Migrating Atlas metadata when upgrading to HDP-3.0+... 3 Overview... 3 Migrate

More information

Hortonworks Data Platform

Hortonworks Data Platform Apache Ambari Views () docs.hortonworks.com : Apache Ambari Views Copyright 2012-2017 Hortonworks, Inc. All rights reserved. The, powered by Apache Hadoop, is a massively scalable and 100% open source

More information

KNIME Extension for Apache Spark Installation Guide

KNIME Extension for Apache Spark Installation Guide Installation Guide KNIME GmbH Version 2.3.0, July 11th, 2018 Table of Contents Introduction............................................................................... 1 Supported Hadoop distributions...........................................................

More information

Hortonworks SmartSense

Hortonworks SmartSense Hortonworks SmartSense Installation (January 8, 2018) docs.hortonworks.com Hortonworks SmartSense: Installation Copyright 2012-2018 Hortonworks, Inc. Some rights reserved. The Hortonworks Data Platform,

More information

EMC ISILON ONEFS WITH HADOOP AND CLOUDERA

EMC ISILON ONEFS WITH HADOOP AND CLOUDERA EMC ISILON ONEFS WITH HADOOP AND CLOUDERA FOR KERBEROS INSTALLATION GUIDE VERSION 1.03 Abstract This guide walks you through the process of installing EMC Isilon OneFS with the Cloudera for Kerberos distribution

More information

Ranger installation in Kerberized Environment

Ranger installation in Kerberized Environment Ranger installation in Kerberized Environment Summary Creating Keytab and principals For Ranger Admin For Ranger Lookup For Ranger Usersync For Ranger Tagsync Installation Steps for Ranger-Admin Installation

More information

Configuring Hadoop Security with Cloudera Manager

Configuring Hadoop Security with Cloudera Manager Configuring Hadoop Security with Cloudera Manager Important Notice (c) 2010-2015 Cloudera, Inc. All rights reserved. Cloudera, the Cloudera logo, Cloudera Impala, and any other product or service names

More information

docs.hortonworks.com

docs.hortonworks.com docs.hortonworks.com Hortonworks Data Platform : Security Administration Tools Guide Copyright 2012-2014 Hortonworks, Inc. Some rights reserved. The Hortonworks Data Platform, powered by Apache Hadoop,

More information

Introduction to Cloudbreak

Introduction to Cloudbreak 2 Introduction to Cloudbreak Date of Publish: 2019-02-06 https://docs.hortonworks.com/ Contents What is Cloudbreak... 3 Primary use cases... 3 Interfaces...3 Core concepts... 4 Architecture... 7 Cloudbreak

More information

HDFS Access Options, Applications

HDFS Access Options, Applications Hadoop Distributed File System (HDFS) access, APIs, applications HDFS Access Options, Applications Able to access/use HDFS via command line Know about available application programming interfaces Example

More information

Hortonworks Data Platform

Hortonworks Data Platform Hortonworks Data Platform Apache Ambari Upgrade (October 30, 2017) docs.hortonworks.com Hortonworks Data Platform: Apache Ambari Upgrade Copyright 2012-2017 Hortonworks, Inc. Some rights reserved. The

More information

Configuring NiFi Authentication and Proxying with Apache Knox

Configuring NiFi Authentication and Proxying with Apache Knox 3 Configuring NiFi Authentication and Proxying with Apache Knox Date of Publish: 2018-08-13 http://docs.hortonworks.com Contents...3 Configuring NiFi for Knox Authentication... 3 Configuring Knox for NiFi...

More information

Hortonworks Data Platform

Hortonworks Data Platform Hortonworks Data Platform Apache Ambari Upgrade for IBM Power Systems (May 17, 2018) docs.hortonworks.com Hortonworks Data Platform: Apache Ambari Upgrade for IBM Power Systems Copyright 2012-2018 Hortonworks,

More information

Document Type: Best Practice

Document Type: Best Practice Global Architecture and Technology Enablement Practice Hadoop with Kerberos Deployment Considerations Document Type: Best Practice Note: The content of this paper refers exclusively to the second maintenance

More information

EMC ISILON ONEFS WITH HADOOP AND HORTONWORKS INSTALLATION GUIDE

EMC ISILON ONEFS WITH HADOOP AND HORTONWORKS INSTALLATION GUIDE EMC ISILON ONEFS WITH HADOOP AND HORTONWORKS INSTALLATION GUIDE VERSION 1.07 Abstract This guide walks you through the process of installing Isilon OneFS with Hadoop for use with the Hortonworks Data Platform

More information

Installing Apache Knox

Installing Apache Knox 3 Installing Apache Knox Date of Publish: 2018-07-15 http://docs.hortonworks.com Contents...3 Install Knox...3 Set Up Knox Proxy... 4 Example: Configure Knox Gateway for YARN UI...6 Example: Configure

More information

Hortonworks Data Platform

Hortonworks Data Platform Hortonworks Data Platform Workflow Management (August 31, 2017) docs.hortonworks.com Hortonworks Data Platform: Workflow Management Copyright 2012-2017 Hortonworks, Inc. Some rights reserved. The Hortonworks

More information

Spotfire Security. Peter McKinnis July 2017

Spotfire Security. Peter McKinnis July 2017 Spotfire Security Peter McKinnis July 2017 Outline Authentication in Spotfire Spotfire Server 7.9 Sites Feature and Authentication Authorization in Spotfire Data Security Spotfire Statistics Services Security

More information

Hortonworks SmartSense

Hortonworks SmartSense Hortonworks SmartSense Installation (April 3, 2017) docs.hortonworks.com Hortonworks SmartSense: Installation Copyright 2012-2017 Hortonworks, Inc. Some rights reserved. The Hortonworks Data Platform,

More information

Apache Ranger User Guide

Apache Ranger User Guide Apache Ranger 0.5 - User Guide USER GUIDE Version : 0.5.0 September 2015 About this document Getting started General Features Login to the system: Log out to the system: Service Manager (Access Manager)

More information

VMware vsphere Big Data Extensions Administrator's and User's Guide

VMware vsphere Big Data Extensions Administrator's and User's Guide VMware vsphere Big Data Extensions Administrator's and User's Guide vsphere Big Data Extensions 1.1 This document supports the version of each product listed and supports all subsequent versions until

More information

Nicolas Williams Staff Engineer Sun Microsystems, Inc.

Nicolas Williams Staff Engineer Sun Microsystems, Inc. Deploying Secure NFS Nicolas Williams Staff Engineer Sun Microsystems, Inc. nicolas.williams@sun.com Page 1 of Secure NFS Background A Brief History Protocol In the beginning, no security AUTH_SYS, AUTH_NONE

More information

Hortonworks Data Platform

Hortonworks Data Platform Hortonworks Data Platform Apache Ambari Administration (March 5, 2018) docs.hortonworks.com Hortonworks Data Platform: Apache Ambari Administration Copyright 2012-2018 Hortonworks, Inc. Some rights reserved.

More information

ambari administration 2 Administering Ambari Date of Publish:

ambari administration 2 Administering Ambari Date of Publish: 2 Administering Ambari Date of Publish: 2018-04-30 http://docs.hortonworks.com Contents ii Contents Introducing Ambari administration... 5 Understanding Ambari terminology...5 Using the Administrator role

More information

Hortonworks Data Platform

Hortonworks Data Platform Hortonworks Data Platform Apache Ambari Security (October 30, 2017) docs.hortonworks.com Hortonworks Data Platform: Apache Ambari Security Copyright 2012-2017 Hortonworks, Inc. All rights reserved. The

More information

How to Run the Big Data Management Utility Update for 10.1

How to Run the Big Data Management Utility Update for 10.1 How to Run the Big Data Management Utility Update for 10.1 2016 Informatica LLC. No part of this document may be reproduced or transmitted in any form, by any means (electronic, photocopying, recording

More information

Hortonworks Data Platform

Hortonworks Data Platform Data Governance () docs.hortonworks.com : Data Governance Copyright 2012-2017 Hortonworks, Inc. Some rights reserved. The, powered by Apache Hadoop, is a massively scalable and 100% open source platform

More information

Upgrading Big Data Management to Version Update 2 for Hortonworks HDP

Upgrading Big Data Management to Version Update 2 for Hortonworks HDP Upgrading Big Data Management to Version 10.1.1 Update 2 for Hortonworks HDP Copyright Informatica LLC 2017. Informatica, the Informatica logo, and Informatica Big Data Management are trademarks or registered

More information

SAS Viya 3.3 Administration: Authentication

SAS Viya 3.3 Administration: Authentication SAS Viya 3.3 Administration: Authentication Authentication: Overview...................................................................... 1 Authentication: How To........................................................................

More information

Single Sign On (SSO) with Polarion 17.3

Single Sign On (SSO) with Polarion 17.3 SIEMENS Single Sign On (SSO) with Polarion 17.3 POL007 17.3 Contents Configuring single sign-on (SSO)......................................... 1-1 Overview...........................................................

More information

6Hadoop Security. Securing Hadoop Cluster Securing Data stored in Cluster Securing Applications running in Cluster

6Hadoop Security. Securing Hadoop Cluster Securing Data stored in Cluster Securing Applications running in Cluster 6Hadoop Security WHAT S IN THIS CHAPTER? Securing Hadoop Cluster Securing Data stored in Cluster Securing Applications running in Cluster Given that Hadoop is used for storing and processing an organization

More information

Using Apache Zeppelin

Using Apache Zeppelin 3 Using Apache Zeppelin Date of Publish: 2018-04-01 http://docs.hortonworks.com Contents Introduction... 3 Launch Zeppelin... 3 Working with Zeppelin Notes... 5 Create and Run a Note...6 Import a Note...7

More information

Installation and Upgrade 1. Installing DataPlane. Date of Publish:

Installation and Upgrade 1. Installing DataPlane. Date of Publish: 1 Installing DataPlane Date of Publish: 2018-08-07 http://docs.hortonworks.com Contents DataPlane Platform support requirements...3 Installation overview...4 Pre-installation tasks... 6 Prepare your clusters...

More information

Hortonworks Data Platform

Hortonworks Data Platform Hortonworks Data Platform Apache Ambari Upgrade (July 15, 2018) docs.hortonworks.com Hortonworks Data Platform: Apache Ambari Upgrade Copyright 2012-2018 Hortonworks, Inc. Some rights reserved. The Hortonworks

More information

Using Apache Phoenix to store and access data

Using Apache Phoenix to store and access data 3 Using Apache Phoenix to store and access data Date of Publish: 2018-07-15 http://docs.hortonworks.com Contents ii Contents What's New in Apache Phoenix...4 Orchestrating SQL and APIs with Apache Phoenix...4

More information

Installation 1. Installing DPS. Date of Publish:

Installation 1. Installing DPS. Date of Publish: 1 Installing DPS Date of Publish: 2018-05-18 http://docs.hortonworks.com Contents DPS Platform support requirements...3 Installation overview...4 Installation prerequisites...5 Setting up the local repository

More information

docs.hortonworks.com

docs.hortonworks.com docs.hortonworks.com : Getting Started Guide Copyright 2012, 2014 Hortonworks, Inc. Some rights reserved. The, powered by Apache Hadoop, is a massively scalable and 100% open source platform for storing,

More information

SAS Data Loader 2.4 for Hadoop

SAS Data Loader 2.4 for Hadoop SAS Data Loader 2.4 for Hadoop vapp Deployment Guide SAS Documentation The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2015. SAS Data Loader 2.4 for Hadoop: vapp Deployment

More information

Novell Kerberos Login Method for NMASTM

Novell Kerberos Login Method for NMASTM Novell Kerberos Login Method for NMASTM 1.0 ADMINISTRATION GUIDE www.novell.com Legal Notices Novell, Inc. makes no representations or warranties with respect to the contents or use of this documentation,

More information

Installing HDF Services on an Existing HDP Cluster

Installing HDF Services on an Existing HDP Cluster 3 Installing HDF Services on an Existing HDP Cluster Date of Publish: 2018-08-13 http://docs.hortonworks.com Contents Upgrade Ambari and HDP...3 Installing Databases...3 Installing MySQL... 3 Configuring

More information

Important Notice Cloudera, Inc. All rights reserved.

Important Notice Cloudera, Inc. All rights reserved. Important Notice 2010-2017 Cloudera, Inc. All rights reserved. Cloudera, the Cloudera logo, and any other product or service names or slogans contained in this document, except as otherwise disclaimed,

More information

Oracle Payment Interface Token Proxy Service Security Guide Release 6.1 E November 2017

Oracle Payment Interface Token Proxy Service Security Guide Release 6.1 E November 2017 Oracle Payment Interface Token Proxy Service Security Guide Release 6.1 E87635-01 November 2017 Copyright 2017, Oracle and/or its affiliates. All rights reserved. This software and related documentation

More information

Hortonworks Hive ODBC Driver

Hortonworks Hive ODBC Driver Hortonworks Hive ODBC Driver User Guide Revised: August 17, 2018 2012-2018 Hortonworks Inc. All Rights Reserved. Parts of this Program and Documentation include proprietary software and content that is

More information

Setting Up Identity Management

Setting Up Identity Management APPENDIX D Setting Up Identity Management To prepare for the RHCSA and RHCE exams, you need to use a server that provides Lightweight Directory Access Protocol (LDAP) and Kerberos services. The configuration

More information

Hortonworks Data Platform

Hortonworks Data Platform Hortonworks Data Platform Apache Zeppelin Component Guide (December 15, 2017) docs.hortonworks.com Hortonworks Data Platform: Apache Zeppelin Component Guide Copyright 2012-2017 Hortonworks, Inc. Some

More information

SAS High-Performance Analytics Infrastructure 2.8

SAS High-Performance Analytics Infrastructure 2.8 SAS High-Performance Analytics Infrastructure 2.8 Installation and Configuration Guide SAS Documentation The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2014. SAS High-Performance

More information

ELASTIC DATA PLATFORM

ELASTIC DATA PLATFORM SERVICE OVERVIEW ELASTIC DATA PLATFORM A scalable and efficient approach to provisioning analytics sandboxes with a data lake ESSENTIALS Powerful: provide read-only data to anyone in the enterprise while

More information

How to Configure Big Data Management 10.1 for MapR 5.1 Security Features

How to Configure Big Data Management 10.1 for MapR 5.1 Security Features How to Configure Big Data Management 10.1 for MapR 5.1 Security Features 2014, 2016 Informatica LLC. No part of this document may be reproduced or transmitted in any form, by any means (electronic, photocopying,

More information

Integrating with Apache Hadoop

Integrating with Apache Hadoop HPE Vertica Analytic Database Software Version: 7.2.x Document Release Date: 10/10/2017 Legal Notices Warranty The only warranties for Hewlett Packard Enterprise products and services are set forth in

More information

Oracle WebLogic Server

Oracle WebLogic Server Oracle WebLogic Server Creating WebLogic Domains Using the Configuration Wizard 10g Release 3 (10.1.3) August 2008 Oracle WebLogic Server Creating WebLogic Domains Using the Configuration Wizard, 10g Release

More information

Installing Apache Ranger KMS

Installing Apache Ranger KMS 3 Installing Apache Ranger KMS Date of Publish: 2018-07-15 http://docs.hortonworks.com Contents Installing the Ranger Key Management Service...3 Install Ranger KMS using Ambari (Kerberized Cluster)...

More information

9.4 Hadoop Configuration Guide for Base SAS. and SAS/ACCESS

9.4 Hadoop Configuration Guide for Base SAS. and SAS/ACCESS SAS 9.4 Hadoop Configuration Guide for Base SAS and SAS/ACCESS Second Edition SAS Documentation The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2015. SAS 9.4 Hadoop

More information

Securing the Oracle BDA - 1

Securing the Oracle BDA - 1 Hello and welcome to this online, self-paced course titled Administering and Managing the Oracle Big Data Appliance (BDA). This course contains several lessons. This lesson is titled Securing the Oracle

More information

IBM BigInsights BigIntegrate and BigQuality: IBM InfoSphere Information Server on Hadoop Deployment and Configuration Guide

IBM BigInsights BigIntegrate and BigQuality: IBM InfoSphere Information Server on Hadoop Deployment and Configuration Guide IBM BigInsights BigIntegrate and BigQuality: IBM InfoSphere Information Server on Hadoop Deployment and Configuration Guide IBM BigInsights BigIntegrate and BigQuality allow for IBM InfoSphere Information

More information

Powerful and Frictionless Storage Administration

Powerful and Frictionless Storage Administration Powerful and Frictionless Storage Administration Configuration Guide 2012-2014 SoftNAS, LLC Table of Contents Overview...3 Server Components...4 Kerberos Authentication...5 Prerequisites...6 Configuration

More information

How to Install and Configure EBF16193 for Hortonworks HDP 2.3 and HotFix 3 Update 2

How to Install and Configure EBF16193 for Hortonworks HDP 2.3 and HotFix 3 Update 2 How to Install and Configure EBF16193 for Hortonworks HDP 2.3 and 9.6.1 HotFix 3 Update 2 1993-2015 Informatica Corporation. No part of this document may be reproduced or transmitted in any form, by any

More information

Administration 1. DLM Administration. Date of Publish:

Administration 1. DLM Administration. Date of Publish: 1 DLM Administration Date of Publish: 2018-07-03 http://docs.hortonworks.com Contents ii Contents Replication Concepts... 4 HDFS cloud replication...4 Hive cloud replication... 4 Cloud replication guidelines

More information

Automation of Rolling Upgrade for Hadoop Cluster without Data Loss and Job Failures. Hiroshi Yamaguchi & Hiroyuki Adachi

Automation of Rolling Upgrade for Hadoop Cluster without Data Loss and Job Failures. Hiroshi Yamaguchi & Hiroyuki Adachi Automation of Rolling Upgrade for Hadoop Cluster without Data Loss and Job Failures Hiroshi Yamaguchi & Hiroyuki Adachi About Us 2 Hiroshi Yamaguchi Hiroyuki Adachi Hadoop DevOps Engineer Hadoop Engineer

More information

Important Notice Cloudera, Inc. All rights reserved.

Important Notice Cloudera, Inc. All rights reserved. Cloudera Security Important Notice 2010-2017 Cloudera, Inc. All rights reserved. Cloudera, the Cloudera logo, and any other product or service names or slogans contained in this document are trademarks

More information

Hortonworks Data Platform

Hortonworks Data Platform Apache Spark Component Guide () docs.hortonworks.com : Apache Spark Component Guide Copyright 2012-2017 Hortonworks, Inc. Some rights reserved. The, powered by Apache Hadoop, is a massively scalable and

More information

Isilon OneFS Authentication, Identity Management, & Authorization

Isilon OneFS Authentication, Identity Management, & Authorization Isilon OneFS Authentication, Identity Management, & Authorization Multi-Protocol Data Access & the Unified Permission Model Abstract This white paper details user and file access management in Isilon OneFS

More information

Release Notes 1. DLM Release Notes. Date of Publish:

Release Notes 1. DLM Release Notes. Date of Publish: 1 DLM Release Notes Date of Publish: 2018-05-18 http://docs.hortonworks.com Contents...3 What s New in this Release...3 Behavioral Changes... 3 Known Issues...3 Fixed Issues...5 This document provides

More information

3.6. How to Use the Reports and Data Warehouse Capabilities of Red Hat Enterprise Virtualization. Last Updated:

3.6. How to Use the Reports and Data Warehouse Capabilities of Red Hat Enterprise Virtualization. Last Updated: Red Hat Enterprise Virtualization 3.6 Reports and Data Warehouse Guide How to Use the Reports and Data Warehouse Capabilities of Red Hat Enterprise Virtualization Last Updated: 2017-09-27 Red Hat Enterprise

More information

Workspace ONE UEM Certificate Authentication for EAS with ADCS. VMware Workspace ONE UEM 1902

Workspace ONE UEM Certificate Authentication for EAS with ADCS. VMware Workspace ONE UEM 1902 Workspace ONE UEM Certificate Authentication for EAS with ADCS VMware Workspace ONE UEM 1902 You can find the most up-to-date technical documentation on the VMware website at: https://docs.vmware.com/

More information

High-Performance Analytics Infrastructure 3.1

High-Performance Analytics Infrastructure 3.1 SAS High-Performance Analytics Infrastructure 3.1 Installation and Configuration Guide SAS Documentation The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2015. SAS High-Performance

More information

Cloudera ODBC Driver for Apache Hive

Cloudera ODBC Driver for Apache Hive Cloudera ODBC Driver for Apache Hive Important Notice 2010-2017 Cloudera, Inc. All rights reserved. Cloudera, the Cloudera logo, and any other product or service names or slogans contained in this document,

More information

SAS 9.4 Hadoop Configuration Guide for Base SAS and SAS/ACCESS, Fourth Edition

SAS 9.4 Hadoop Configuration Guide for Base SAS and SAS/ACCESS, Fourth Edition SAS 9.4 Hadoop Configuration Guide for Base SAS and SAS/ACCESS, Fourth Edition SAS Documentation August 31, 2017 The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2016.

More information

SAS High-Performance Analytics Infrastructure 3.5: Installation and Configuration Guide

SAS High-Performance Analytics Infrastructure 3.5: Installation and Configuration Guide SAS High-Performance Analytics Infrastructure 3.5: Installation and Configuration Guide SAS Documentation The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2016. SAS

More information

SAS High-Performance Analytics Infrastructure 3.6: Installation and Configuration Guide

SAS High-Performance Analytics Infrastructure 3.6: Installation and Configuration Guide SAS High-Performance Analytics Infrastructure 3.6: Installation and Configuration Guide SAS Documentation June 2017 The correct bibliographic citation for this manual is as follows: SAS Institute Inc.

More information

Hands-on Exercise Hadoop

Hands-on Exercise Hadoop Department of Economics and Business Administration Chair of Business Information Systems I Prof. Dr. Barbara Dinter Big Data Management Hands-on Exercise Hadoop Building and Testing a Hadoop Cluster by

More information

Server Installation Guide

Server Installation Guide Server Installation Guide Server Installation Guide Legal notice Copyright 2018 LAVASTORM ANALYTICS, INC. ALL RIGHTS RESERVED. THIS DOCUMENT OR PARTS HEREOF MAY NOT BE REPRODUCED OR DISTRIBUTED IN ANY

More information

Configuring and Deploying Hadoop Cluster Deployment Templates

Configuring and Deploying Hadoop Cluster Deployment Templates Configuring and Deploying Hadoop Cluster Deployment Templates This chapter contains the following sections: Hadoop Cluster Profile Templates, on page 1 Creating a Hadoop Cluster Profile Template, on page

More information

Installing an HDF cluster

Installing an HDF cluster 3 Installing an HDF cluster Date of Publish: 2018-08-13 http://docs.hortonworks.com Contents Installing Ambari...3 Installing Databases...3 Installing MySQL... 3 Configuring SAM and Schema Registry Metadata

More information

VMware Workspace ONE Quick Configuration Guide. VMware AirWatch 9.1

VMware Workspace ONE Quick Configuration Guide. VMware AirWatch 9.1 VMware Workspace ONE Quick Configuration Guide VMware AirWatch 9.1 A P R I L 2 0 1 7 V 2 Revision Table The following table lists revisions to this guide since the April 2017 release Date April 2017 June

More information

BlackBerry UEM Configuration Guide

BlackBerry UEM Configuration Guide BlackBerry UEM Configuration Guide 12.9 2018-11-05Z 2 Contents Getting started... 7 Configuring BlackBerry UEM for the first time... 7 Configuration tasks for managing BlackBerry OS devices... 9 Administrator

More information

Installation and Configuration Guide Simba Technologies Inc.

Installation and Configuration Guide Simba Technologies Inc. Simba Hive ODBC Driver with SQL Connector Installation and Configuration Guide Simba Technologies Inc. Version 2.6.1 August 3, 2018 Copyright 2018 Simba Technologies Inc. All Rights Reserved. Information

More information

Configuration Guide. BlackBerry UEM. Version 12.9

Configuration Guide. BlackBerry UEM. Version 12.9 Configuration Guide BlackBerry UEM Version 12.9 Published: 2018-07-16 SWD-20180713083904821 Contents About this guide... 8 Getting started... 9 Configuring BlackBerry UEM for the first time...9 Configuration

More information

Entrust Connector (econnector) Venafi Trust Protection Platform

Entrust Connector (econnector) Venafi Trust Protection Platform Entrust Connector (econnector) For Venafi Trust Protection Platform Installation and Configuration Guide Version 1.0.5 DATE: 17 November 2017 VERSION: 1.0.5 Copyright 2017. All rights reserved Table of

More information

setup cross realm trust between two MIT KDC to access and copy data of one cluster from another if the cross realm trust is setup correctly.

setup cross realm trust between two MIT KDC to access and copy data of one cluster from another if the cross realm trust is setup correctly. ####################################################### # How to setup cross realm trust between two MIT KDC ####################################################### setup cross realm trust between two

More information

Subversion Plugin HTTPS Kerberos authentication

Subversion Plugin HTTPS Kerberos authentication Subversion Plugin HTTPS Kerberos authentication Introduction Prerequisites Configure the Oracle JRE with Java Cryptography Extension (JCE) Server certificates Prepare and test the domain account Linux

More information

Using the Cisco ACE Application Control Engine Application Switches with the Cisco ACE XML Gateway

Using the Cisco ACE Application Control Engine Application Switches with the Cisco ACE XML Gateway Using the Cisco ACE Application Control Engine Application Switches with the Cisco ACE XML Gateway Applying Application Delivery Technology to Web Services Overview The Cisco ACE XML Gateway is the newest

More information

Designing Windows Server 2008 Network and Applications Infrastructure

Designing Windows Server 2008 Network and Applications Infrastructure Designing Windows Server 2008 Network and Applications Infrastructure Course No. 6435B - 5 Days Instructor-led, Hands-on Introduction This five-day course will provide students with an understanding of

More information

Policy Manager for IBM WebSphere DataPower 7.2: Configuration Guide

Policy Manager for IBM WebSphere DataPower 7.2: Configuration Guide Policy Manager for IBM WebSphere DataPower 7.2: Configuration Guide Policy Manager for IBM WebSphere DataPower Configuration Guide SOAPMDP_Config_7.2.0 Copyright Copyright 2015 SOA Software, Inc. All rights

More information

Red Hat Quay 2.9 Deploy Red Hat Quay - Basic

Red Hat Quay 2.9 Deploy Red Hat Quay - Basic Red Hat Quay 2.9 Deploy Red Hat Quay - Basic Deploy Red Hat Quay Last Updated: 2018-09-14 Red Hat Quay 2.9 Deploy Red Hat Quay - Basic Deploy Red Hat Quay Legal Notice Copyright 2018 Red Hat, Inc. The

More information