iway Big Data Integrator New Features Bulletin and Release Notes

Similar documents
iway iway Big Data Integrator New Features Bulletin and Release Notes Version DN

iway iway Big Data Integrator New Features Bulletin and Release Notes Version DN

WebFOCUS RStat Release Guide Version 3.0

Using the DataDirect ODBC Driver for Hyperstage

Server Release Notes WebFOCUS Reporting Server Release 8203 DataMigrator Server Release 7708

iway iway Big Data Integrator User s Guide Version DN

Source Management (Version Control) Installation and Configuration Guide. Version 8.0 and Higher

Mobile Faves for Android User's Manual Version 3 Release 2.1 and Higher

iway Integration Tools Getting Started Guide Version 8.0 and Higher

WebFOCUS Business User Edition Release Guide Release 8.2 Version 01

Troubleshooting Release 8.2 Version 01M

WebFOCUS Open Portal Services Administration Guide. Release 8.0 Version 09

iway iway Big Data Integrator Getting Started Lab Version DN

Server Release Notes. WebFOCUS Reporting Server Release 82. DataMigrator Server Release 7707 DN

WebFOCUS Release Notes Release 8.2 Version 04

WebFOCUS Narrative Charts Installation and Configuration Release 8.2 Version 04

Omni-Gen Consumption View User s Guide. Version 3.8

Managed Reporting Release 8.2 Version 01

Troubleshooting Release 8.2 Version 02

Mobile Faves for Android User's Manual Version 3.1

Omni-Gen API Services Reference Guide. Version 3.6

iway Omni-Payer Management Central User s Guide Version DN

Managed Reporting Release 8.2 Version 03

Creating WebFOCUS Infographics Release 8.2 Version 04

App Studio for the Developer Studio User

Mobile Faves Branding Package Developer s Guide

Innovatus Technologies

IBM DB2 Web Query for IBM i. Version 2 Release 2

Migration Release 8.0 Version 09

IBM DB2 Web Query for IBM i. Version 2 Release 2

DB2 Web Query Active Technologies for Mobile Web Apps

SAS Data Loader 2.4 for Hadoop

Using WebFOCUS Designer Release 8.2 Version 03

Omni Console User s Guide. Version 3.6.3

Data Quality Workbench User s Guide. Version 3.2 and Higher

Migration Release 8.2 Version 04

Hortonworks Data Platform

WebFOCUS Business User Edition Release 8.2 Version 02

iway Software Development Kit User's Guide Version 8.0 and Higher

WebFOCUS Managed Reporting End User's Manual

WebFOCUS Business User Edition Release 8.2 Version 01M

Overview. : Cloudera Data Analyst Training. Course Outline :: Cloudera Data Analyst Training::

Working With Data Release 8.2 Version 01M

WebFOCUS App Studio Installation and Configuration Guide Release 8.2 Version 04

WebFOCUS App Studio Installation and Configuration Guide Release 8.2 Version 02

Blended Learning Outline: Developer Training for Apache Spark and Hadoop (180404a)

Big Data Hadoop Developer Course Content. Big Data Hadoop Developer - The Complete Course Course Duration: 45 Hours

Hadoop. Course Duration: 25 days (60 hours duration). Bigdata Fundamentals. Day1: (2hours)

Blended Learning Outline: Cloudera Data Analyst Training (171219a)

Talend Big Data Sandbox. Big Data Insights Cookbook

How to Run the Big Data Management Utility Update for 10.1

Oracle Big Data Manager User s Guide. For Oracle Big Data Appliance

Getting Started With the Responsive Design Framework Add-on for WebFOCUS App Studio Release 8.1 Version 05

We are ready to serve Latest Testing Trends, Are you ready to learn?? New Batches Info

Oracle Cloud Using Oracle Big Data Manager. Release

Big Data Hadoop Course Content

Oracle Cloud Using Oracle Big Data Manager. Release

CERTIFICATE IN SOFTWARE DEVELOPMENT LIFE CYCLE IN BIG DATA AND BUSINESS INTELLIGENCE (SDLC-BD & BI)

Hadoop & Big Data Analytics Complete Practical & Real-time Training

Certified Big Data and Hadoop Course Curriculum

Overview. Prerequisites. Course Outline. Course Outline :: Apache Spark Development::

DB2 Web Query New Features. July 2010 Hotfix

Certified Big Data Hadoop and Spark Scala Course Curriculum

WebFOCUS Release Notes Release 8.2 Version 02M

Active Technologies User's Guide Release 8.2 Version 01M

PUBLIC SAP Vora Sizing Guide

iway Release Notes Version 8.0 Service Manager (SM) DN

Enterprise Data Catalog Fixed Limitations ( Update 1)

Oracle Big Data SQL. Release 3.2. Rich SQL Processing on All Data

Informatica PowerExchange for Microsoft Azure Blob Storage 10.2 HotFix 1. User Guide

Security and Performance advances with Oracle Big Data SQL

Important Notice Cloudera, Inc. All rights reserved.

Cloudera Manager Quick Start Guide

Delving Deep into Hadoop Course Contents Introduction to Hadoop and Architecture

SAS Data Loader 2.4 for Hadoop: User s Guide

Configuring Intelligent Streaming 10.2 For Kafka on MapR

VMware vsphere Big Data Extensions Administrator's and User's Guide

Cloudera Connector for Netezza

Oracle Big Data Fundamentals Ed 2

Pre-Installation Tasks Before you apply the update, shut down the Informatica domain and perform the pre-installation tasks.

iway BEA WebLogic Solutions Guide for iway Version 5.5

Is NiFi compatible with Cloudera, Map R, Hortonworks, EMR, and vanilla distributions?

Quick Install for Amazon EMR

Big Data Analytics using Apache Hadoop and Spark with Scala

New in This Version. Numeric Filtergram

The Hadoop Ecosystem. EECS 4415 Big Data Systems. Tilemachos Pechlivanoglou

VMware vsphere Big Data Extensions Command-Line Interface Guide

HADOOP COURSE CONTENT (HADOOP-1.X, 2.X & 3.X) (Development, Administration & REAL TIME Projects Implementation)

Exam Questions

Introduction to BigData, Hadoop:-

iway iway Java Adapter for Mainframe Introduction

Want to read more? You can buy this book at oreilly.com in print and ebook format. Buy 2 books, get the 3rd FREE!

Hadoop. Introduction to BIGDATA and HADOOP

Big Data Integrator Platform Platform Architecture and Features Dr. Hajira Jabeen Technical Team Leader-BDE University of Bonn

Talend Big Data Sandbox. Big Data Insights Cookbook

See Types of Data Supported for information about the types of files that you can import into Datameer.

Release Notes for Oracle GoldenGate for Big Data 12c ( )

WebFOCUS Adapter for Geographic Information Systems Getting Started Release 8.2 Version 02

arxiv: v1 [cs.dc] 20 Aug 2015

Microsoft. Exam Questions Perform Data Engineering on Microsoft Azure HDInsight (beta) Version:Demo

Transcription:

iway Big Data Integrator New Features Bulletin and Release Notes Version 1.5.2 DN3502232.0717

Active Technologies, EDA, EDA/SQL, FIDEL, FOCUS, Information Builders, the Information Builders logo, iway, iway Software, Parlay, PC/FOCUS, RStat, Table Talk, Web390, WebFOCUS, WebFOCUS Active Technologies, and WebFOCUS Magnify are registered trademarks, and DataMigrator and Hyperstage are trademarks of Information Builders, Inc. Adobe, the Adobe logo, Acrobat, Adobe Reader, Flash, Adobe Flash Builder, Flex, and PostScript are either registered trademarks or trademarks of Adobe Systems Incorporated in the United States and/or other countries. Due to the nature of this material, this document refers to numerous hardware and software products by their trademarks. In most, if not all cases, these designations are claimed as trademarks or registered trademarks by their respective companies. It is not this publisher's intent to use any of these names generically. The reader is therefore cautioned to investigate all claimed trademark rights before using any of these names other than to refer to the product described. Copyright 2017, by Information Builders, Inc. and iway Software. All rights reserved. Patent Pending. This manual, or parts thereof, may not be reproduced in any form without the written permission of Information Builders, Inc.

Chapter1 iway Big Data Integrator Version 1.5.2 This document describes new features and provides release notes for iway Big Data Integrator (ibdi) version 1.5.2. It is intended for all levels of users, including system integrators, application developers, and administrators. In this chapter: New Features Release Notes New Features This section provides a summary of the new features for iway Big Data Integrator (ibdi) version 1.5.2. Smaller Cluster Sizes (BDI-504) ibdi can now run in smaller cluster sizes. Non-Pipeline Objects (BDI-513) Non-pipeline objects can now specify host name and port values. Release Notes This section provides release notes for iway Big Data Integrator (ibdi) version 1.5.2 and version 1.5.1. Important Considerations This section describes important considerations that you should be aware when using iway Big Data Integrator (ibdi) version 1.5.2. iway Big Data Integrator New Features Bulletin and Release Notes 3

Release Notes JDBC Drivers JDBC drivers are not shipped with ibdi and must be obtained from a relevant JDBC vendor (for example, Cloudera and Simba). Depending on your usage requirements, there are several open source Hive JDBC drivers available online. For additional security and availability options, commercial drivers are also available. Each driver includes a list of dependencies for the specific driver. Consult the driver provider for the list of.jar files that must be used with the specific driver version. Data Definition Language (DDL) Data Definition Language (DDL) generated by ibdi is specific to the JDBC driver being used. DDL generated for one database type (for example, MySQL) cannot be applied in another domain (for example, Hive). Transformer In a pipeline, the Transformer is only supported for Hive table objects. Known Issues in Version 1.5.2 This section describes known issues for iway Big Data Integrator (ibdi) in version 1.5.2. Defining Run Configurations for a Flume An endless loop occurs when using the iway Big Data Integrator Deploy option for a Flume run configuration. Resolution: Always use the iway Big Data Integrator Publish option when creating run configurations for streaming and event options. Thread Creation Error When running the ibdi Client (Eclipse design time framework/workbench) in a memory constrained environment (for example, a vendor sandbox), the following error message may appear: cannot create new native thread Do not close the ibdi workbench, but click through the error dialog and continue. Resolution: Increase the virtual machine memory limits. Hive Connection Object Entries cannot be edited directly when using a Hive connection object. 4 Information Builders

1. iway Big Data Integrator Version 1.5.2 Resolution: Create another Hive target entry and access the profile through the new entry. XML and JSON Deserializers The current implementation requires flat input only. Expression Field or a Constant Field in a Transform or Map An expression field or a constant field in a transform or map cannot be renamed, even though the field appears to be renamed in the user interface. Pipelines The known issues described in this section apply to pipelines in ibdi version 1.5.2. Pipeline Transformer Tool: Define Mappings Dialog (BDI-554) The Define Mappings dialog currently allows the following incorrect behavior to occur: Right-clicking and selecting Add Column from the context menu. A NULL column is created in the Hive HQL, which invalidates the transform. Do not add these columns. Clicking on a column to rename its current label/title. Although you may specify a new name for a column using this action, the renamed value for the column is ignored. Compute Type: Save The pipeline Compute type Save generates a class not found exception. Resolution: If you must use Save as a Compute type, then compose a Hive Query Language Statement using syntax similar to the following: CREATE TABLE [IF NOT EXISTS] [db_name.]table_name LIKE existing_table_ [LOCATION hdfs_path]; The complete Hive Data Definition Language (DDL) syntax can be found in the Apache Hive Language/DDL documentation. iway Big Data Integrator New Features Bulletin and Release Notes 5

Release Notes Compute Type: Coalesce The pipeline Compute type Coalesce can only be used when the number of partitions is greater than the value entered for Coalesce. For example, if you have 20 partitions, and enter a value of 30 for Coalesce, then the Spark runtime will generate an exception. Hive Source to Hive Compute (BDI-553) A pipeline using a Hive source to Hive compute generates a null pointer exception (Failed to create the part s controls). Resolution: Do not use back to back transforms in a pipeline. Use Hive and an HQL object. Data Serialization in Pipeline Components Data format types in a pipeline source do not require a schema if they are of type text. Otherwise, a schema is required to deserialize the source. If a CSV file is used where the first line is a header describing the fields, then the infer schema can be used. However, the preference is still to have an Avro schema describing the structure, fields, and type of data. Some pipeline components require deserialization from specific formats (for example, pipeline source of type HDFS, Kafka, NiFi, or Stream) or to specific formats, known as serialization (for example, a pipeline compute type of Save or a pipeline target type of HDFS). These components provide a drop-down list allowing you to select the data format type and an edit field that allows you to specify a schema name. Source requirements: If the data format type is a streaming type (for example, Kafka, Kafka Direct, NiFi, or Stream), then a schema must be supplied in Avro format describing the stream. The schema should have a flat structure without any nested objects. Kafka Topics Kafka topics must be pre-configured for use as pipeline target producers. Pipeline Target for HDFS Output Do not select CSV as an option when specifying a pipeline target for HDFS output. Select the HDFS CSV option to write CSV to HDFS. Pipeline Target Data File Formats Avro will be saved as Avro, and Orc will be saved as orc. By default, all other format types will be saved as Parquet files. Use HDFS CSV to save to HDFS CSV. 6 Information Builders

1. iway Big Data Integrator Version 1.5.2 Resolved Issues in Version 1.5.2 This section describes resolved issues for iway Big Data Integrator (ibdi) in version 1.5.2. Sending Files to the Edge Node Host Generates an Error Message on CentOS or Ubuntu (BDI-552) An error in the Secure Copy command when a timestamp is created generates an error message in the Bash shell of the host (mtime.sec not present) when using the following operating systems: Ubuntu version 16 and higher CentOS version 7 and higher Resolution: If you are using an earlier version of ibdi (1.5.1 or 1.5.0), then you must install ibdi version 1.5.2 Client (Eclipse design time framework/workbench) and provision a server edge node with ibdi version 1.5.2. Known Issues in Version 1.5.1 This section describes known issues for iway Big Data Integrator (ibdi) in version 1.5.1. Issue Number Description Resolution BDI-528 BDI-532 BDI-535 BDI-536 BDI-539 Deploy DDL may not create tables for some database types. Sqoop failure to create field when type is tinyint. Transformer functions may lack descriptions. Nifi missing.jar files exception. Hive Source in Pipeline generates a No Such Method error. Database syntax is driver dependent, and the generic syntax created by the Eclipse data tools may need editing before submission. For more information, see section 27.2.5 in the Sqoop User's Guide. For more information, click here to view related information on functions. Nifi use is considered experimental: Advanced deployment necessary, see add on document for Nifi. Do not copy JDBC drivers to the JDBC folder of the ibdi Edge node location. iway Big Data Integrator New Features Bulletin and Release Notes 7

Release Notes Issue Number Description Resolution BDI-540 BDI-541 BDI-542 BDI-543 BDI-544 Data Source Explorer - Hive JDBC cannot edit driver after creation. Pipeline Hive to HDFS nosuchelementexception:key not found:header. Pipeline Rename Null Pointer Exception in run configurations. DDL Create Table fails to create table under Hive database Schema. DDL - Create Schema does not create database schema for mysql. Create a new driver as a copy of the existing driver and modify the properties. Do not use the HDFS with CSV selection option as the pipeline target. Use HDFS CSV as the pipeline target. CSV file format as source in pipeline check the field header as on or off depending on the presence of a header line in the source file. When renaming a pipeline, any run configurations for that pipeline must be removed before renaming. Otherwise, when opening the Run Configurations dialog, a Null Pointer Exception will be generated for trying to read a nonexisting pipeline. Do not use upper case characters and quoted identifiers. For more information, see the Create Table topic in the Apache Hive DDL Language Manual. The JDBC driver for mysql is database dependent, so the database schema must be created in the mysql workbench and not in the DDL execution engine in ibdi. BDI-545 Pipeline Hive Transformer. Rename target column fails to rename column. As a workaround, use the Big Data Mapper to rename table columns. 8 Information Builders

1. iway Big Data Integrator Version 1.5.2 Issue Number Description Resolution BDI-547 Flume used with Run Configuration Deploy option results in an endless loop. Always use Publish for a Flume. The following table lists and describes known issues that are related to Virtual Machines when using iway Big Data Integrator (ibdi) version 1.5.1. Issue Number Description Resolution BDI-534 CDH Hadoop Explorer - NullPointer Exception cannot create thread. Do not close the workbench, click through dialog boxes, thread model related. BDI-537 BDI-538 BDI-523 BDI-524 CDH Sqoop import cannot load Postgres driver in Sqoop. CDH Sqoop Export IoException when destination database is not on the Virtual Machine (VM). HDW JDBC to Hive cannot open connection after close and reopen Workbench. HDW Copying JDBC drivers results in no class definition found. Postgres is not installed on CDH. CDH limits outbound connections. Do not use this option. This is related to the JDBC driver. Do not copy JDBC drivers to the JDBC folder of the ibdi Edge node location. Note: In this table: CDH - Cloudera HDW - HortonWorks iway Big Data Integrator New Features Bulletin and Release Notes 9

Release Notes Run Configurations The following types of run configurations are available: iway Big Data Deploy iway Big Data Pipeline IWay Big Data Publish The Pipeline run configuration is for all pipelines. You can set the Deploy or Publish option for each pipeline inside the configuration. The non-pipeline objects have explicit Deploy or Publish configurations. The Deploy run configuration uploads the objects to the Edge node and runs the configuration immediately, returning the results to the console in the ibdi workbench. Use Deploy when running simple objects with smaller amounts of data that can be run in the ibdi console The Publish run configuration uploads the objects to the Edge node. A user logs in to the Edge node via SSH or another tool, navigates to the folder where the configuration is located, and runs the shell script that triggers the job. The Publish run configuration is mandatory for Flume, Kafka, Kafka Direct, and Stream. It is recommended when using HDFS with large results. Warning: Do not use the Deploy run configuration with Flume. For more information, see BDI 547. Using iway Big Data Integrator on Vendor Provided Sandbox Virtual Machines iway Big Data Integrator (ibdi) version 1.5.2 supports the following versions of common Hadoop software: Sqoop version 1.4.6 Flume version 1.6.0 Spark version 1.6.0 Kafka version 2.10-8.2.1 ibdi Client requires Java version 8 to support the Pipeline and Transformer features. ibdi is designed to run on Hadoop Clusters, not single node machines. However, running in a small Virtual Machine (VM) footprint is possible by adhering to the following guidelines: Please check the terms of service (TOS) of the VM vendor to see if installing third-party software is permitted. Most do not permit this without explicit permission. Installing other software or modifying the VM can violate the terms with the vendor. Check before you install. 10 Information Builders

1. iway Big Data Integrator Version 1.5.2 The sandbox VMs (Cloudera, HortonWorks) have statically assigned host names and will reshuffle assigned IP addresses to match their internal configuration. If these are tampered with, then the sandbox will usually not run. Check the vendor of the VM hostware (Oracle VirtualBox, VMware, or other) on their networking rules. The usual network setup for a sandbox VM will have one network port as a Network Address Translation (NAT) port, and you define a second port as host only. The ibdi communication between the local computer and the VM is through the host only port. The VMs usually have restrictions on outbound communication, so these modes should be sufficient. Assign sufficient memory to the VM. 8GB of memory is the smallest configuration possible under most circumstances. However, 10GB, 12GB, or higher provides better results. Cloudera Sandbox does not have Java version 8 pre-installed. If you intend to install ibdi on the Cloudera VM, then Java version 8 must be installed. Cloudera Sandbox does not have Kafka pre-installed. If you intend to test Kafka, then you will need to install Kafka. The default implementation of Spark does not have a Hive configuration, and the Spark memory defaults must be adjusted. Install ibdi on the image and run the product from the image. Doing so may void your Cloudera terms of service. HortonWorks Sandbox runs inside a Docker image. The usual SSH port number 22 opens the Docker image, and port 2222 opens the actual sandbox. Before using SSH, you must set the password through the Hortonworks console or web console. If you are using a database such as mysql, then you must allow Docker to open the default mysql port number 3306 (or other database port numbers). There may also be permission issues if using mysql or other databases on the local machine. Changing these options may void your HortonWorks terms of service. Install ibdi on your local machine and communicate with Hortonworks through the network address and IP port. MapR MapR is not supported with ibdi at this time. This is currently under research. iway Big Data Integrator New Features Bulletin and Release Notes 11

Release Notes 12 Information Builders

Feedback Customer success is our top priority. Connect with us today! Information Builders Technical Content Management team is comprised of many talented individuals who work together to design and deliver quality technical documentation products. Your feedback supports our ongoing efforts! You can also preview new innovations to get an early look at new content products and services. Your participation helps us create great experiences for every customer. To send us feedback or make a connection, contact Sarah Buccellato, Technical Editor, Technical Content Management at Sarah_Buccellato@ibi.com. To request permission to repurpose copyrighted material, please contact Frances Gambino, Vice President, Technical Content Management at Frances_Gambino@ibi.com.

Creating Reports With WebFOCUS Language iway Big Data Integrator New Features Bulletin and Release Notes Version 7 Release 6 Version 1.5.2 Information Builders Two Penn Plaza New York, NY 10121-2898 Printed on recycled paper in the U.S.A.