How to integrate data into Tableau

Similar documents
The Truth About Test Data Management & Its Impact on Agile Development

Fast Innovation requires Fast IT

JAVASCRIPT CHARTING. Scaling for the Enterprise with Metric Insights Copyright Metric insights, Inc.

Shine a Light on Dark Data with Vertica Flex Tables

Progress DataDirect For Business Intelligence And Analytics Vendors

Strategic Briefing Paper Big Data

Low Friction Data Warehousing WITH PERSPECTIVE ILM DATA GOVERNOR

PERSPECTIVE. Data Virtualization A Potential Antidote for Big Data Growing Pains. Abstract

When, Where & Why to Use NoSQL?

Real Time for Big Data: The Next Age of Data Management. Talksum, Inc. Talksum, Inc. 582 Market Street, Suite 1902, San Francisco, CA 94104

Improving the ROI of Your Data Warehouse

Drawing the Big Picture

Why you should design your data hub top-down vs. bottom-up

CONSOLIDATING RISK MANAGEMENT AND REGULATORY COMPLIANCE APPLICATIONS USING A UNIFIED DATA PLATFORM

Introduction to Data Science

Composite Software Data Virtualization The Five Most Popular Uses of Data Virtualization

An Introduction to Big Data Formats

The Data Explosion. A Guide to Oracle s Data-Management Cloud Services

Data Analytics at Logitech Snowflake + Tableau = #Winning

Massive Scalability With InterSystems IRIS Data Platform

FIVE BEST PRACTICES FOR ENSURING A SUCCESSFUL SQL SERVER MIGRATION

In-Memory Analytics with EXASOL and KNIME //

From Single Purpose to Multi Purpose Data Lakes. Thomas Niewel Technical Sales Director DACH Denodo Technologies March, 2019

Data 101 Which DB, When. Joe Yong Azure SQL Data Warehouse, Program Management Microsoft Corp.

Xcelerated Business Insights (xbi): Going beyond business intelligence to drive information value

The Role of Converged and Hyper-converged Infrastructure in IT Transformation

A primer to SQL Server 2012

SOFTWARE-DEFINED NETWORKING WHAT IT IS, AND WHY IT MATTERS

Incremental Updates VS Full Reload

Best Practices to Transition to the Cloud. Five ways to improve IT agility and speed development by adopting a Cloud DevOps approach

The #1 Key to Removing the Chaos. in Modern Analytical Environments

ELTMaestro for Spark: Data integration on clusters

Building Self-Service BI Solutions with Power Query. Written By: Devin

QLogic 16Gb Gen 5 Fibre Channel for Database and Business Analytics

Big Data with Hadoop Ecosystem

Azure Integration Services

Syncsort DMX-h. Simplifying Big Data Integration. Goals of the Modern Data Architecture SOLUTION SHEET

ATA DRIVEN GLOBAL VISION CLOUD PLATFORM STRATEG N POWERFUL RELEVANT PERFORMANCE SOLUTION CLO IRTUAL BIG DATA SOLUTION ROI FLEXIBLE DATA DRIVEN V

QLogic/Lenovo 16Gb Gen 5 Fibre Channel for Database and Business Analytics

How to Accelerate Merger and Acquisition Synergies

WHITEPAPER. MemSQL Enterprise Feature List

Taming Structured And Unstructured Data With SAP HANA Running On VCE Vblock Systems

Hype Cycle for Data Warehousing, 2003

Introduction to Big-Data

NoSQL database and its business applications

IBM dashdb Local. Using a software-defined environment in a private cloud to enable hybrid data warehousing. Evolving the data warehouse

I D C T E C H N O L O G Y S P O T L I G H T. V i r t u a l and Cloud D a t a Center Management

Azure Data Factory. Data Integration in the Cloud

Introduction to K2View Fabric

SAP Agile Data Preparation Simplify the Way You Shape Data PUBLIC

The Complete Guide to Data Integration 2017

Was ist dran an einer spezialisierten Data Warehousing platform?

Case Study: Tata Communications Delivering a Truly Interactive Business Intelligence Experience on a Large Multi-Tenant Hadoop Cluster

ENTERPRISE DATA STRATEGY IN THE HEALTHCARE LANDSCAPE

ELTMaestro for RedShift: ELT in the Cloud

Bisnode View Why is it so damn hard to piece together information across the enterprise?

An Information Asset Hub. How to Effectively Share Your Data

Next Generation Backup: Better ways to deal with rapid data growth and aging tape infrastructures

Realizing the Full Potential of MDM 1

Transforming IT: From Silos To Services

Cisco ACI App Center. One Platform, Many Applications. Overview

The Emerging Data Lake IT Strategy

Stages of Data Processing

USERS CONFERENCE Copyright 2016 OSIsoft, LLC

In-Memory Computing EXASOL Evaluation

UNLEASHING THE VALUE OF THE TERADATA UNIFIED DATA ARCHITECTURE WITH ALTERYX

WHITE PAPER AUTHENTICATION YOUR WAY SECURING ACCESS IN A CHANGING WORLD

Two Success Stories - Optimised Real-Time Reporting with BI Apps

Data 101 Which DB, When Joe Yong Sr. Program Manager Microsoft Corp.

Evaluation Guide for ASP.NET Web CMS and Experience Platforms

From the past to the future: How to make the move from ISDN to SIP

Hybrid Data Platform

The Business Value of Metadata for Data Governance: The Challenge of Integrating Packaged Applications

INTRODUCTION. Chris Claterbos, Vlamis Software Solutions, Inc. REVIEW OF ARCHITECTURE

Automate Transform Analyze

Modern Data Warehouse The New Approach to Azure BI

MAPR DATA GOVERNANCE WITHOUT COMPROMISE

THE RISE OF. The Disruptive Data Warehouse

Transform to Your Cloud

Lambda Architecture for Batch and Stream Processing. October 2018

Title: Episode 11 - Walking through the Rapid Business Warehouse at TOMS Shoes (Duration: 18:10)

Oracle Exadata: Strategy and Roadmap

Ultra-Low Latency Down to Microseconds SSDs Make It. Possible

Changing the way companies run their data centers

VOLTDB + HP VERTICA. page

TIBCO Data Virtualization for the Energy Industry

TECHNOLOGY WHITE PAPER. Java for the Real Time Business

The Value of Data Modeling for the Data-Driven Enterprise

Abstract. The Challenges. ESG Lab Review InterSystems IRIS Data Platform: A Unified, Efficient Data Platform for Fast Business Insight

Why Converged Infrastructure?

IBM Software IBM InfoSphere Information Server for Data Quality

Building a Data Strategy for a Digital World

Modernizing Business Intelligence and Analytics

Accelerate your SAS analytics to take the gold

SQL Maestro and the ELT Paradigm Shift

Asanka Padmakumara. ETL 2.0: Data Engineering with Azure Databricks

Top Five Reasons for Data Warehouse Modernization Philip Russom

WHAT CIOs NEED TO KNOW TO CAPITALIZE ON HYBRID CLOUD

How a Federated Identity Service Turns Identity into a Business Enabler, Not an IT Bottleneck

The future of database technology is in the clouds

Transcription:

1 How to integrate data into Tableau a comparison of 3 approaches: ETL, Tableau self-service and WHITE PAPER

WHITE PAPER 2 data How to integrate data into Tableau a comparison of 3 es: ETL, Tableau self-service and and data The era of big data is upon us, and with it the dawn of a new industrial revolution. The theoretical benefits of all of this data are tantalizing. But, that s just it. Until data can be unlocked, those benefits remain elusive. While data is more important than ever, it s also more complex, and there s more of it far more.

WHITE PAPER 3 data and In order to yield insight, data must be integrated. The goal of data is to gather data from a variety of different sources, combine it, and present it as a unified whole. However, synchronizing huge quantities of variable, heterogeneous data from disparate, incompatible sources across an enterprise poses significant s. While integrating data has never been easy, the difficulty is only increasing with the proliferation of data sources, types, and stores. In addition to structured data, enormous amounts of raw data are being captured. Much of this data, such as JSON documents or social media, have no schema at all. Combining all of this data to produce meaningful insights is no small task. At the same time, the pace of doing business has accelerated considerably. Business wants to integrate data into their dayto-day operations to help them make important decisions and increase profits. Competitive pressures and new sources of data are creating new requirements, and business users are demanding the ability to answer their questions quickly and easily. Business needs to know if an idea is viable immediately and expects IT to respond with a prototype that can be tweaked accordingly, in the moment. Slow, rigid systems are out of the question for these users and the IT teams that support them. In this blog series, we will explore data, with a special focus on integrating data into Tableau, an interactive data visualization product focused on business intelligence. When it comes to integrating data, business intelligence managers can choose from multiple data approaches. In blog 2 of our series, we ll explore three of them, with special attention paid to how they work with popular BI tool Tableau. Specifically, we ll focus on ETL, Tableau Data Blending, and logical data warehouse/tableau.

WHITE PAPER 4 data and ETL (AND THE TRADITIONAL DATA WAREHOUSE) Tried and true ETL (extract, transform, load) tools can be used to move large amounts of data in a batch-oriented manner. However, when it comes to getting value out of data, these tools pose significant s. Because they require comprehensive knowledge of each operational database or application involved and increasingly complex custom s, ETL-based projects tend to experience a high failure rate. Architecturally rigid in nature, even small changes trigger large and unpredictable impacts. To avoid this, great care must be taken to conceptualize the database and determine requirements. By the time business is able to see the results of the effort, months have passed, and requirements have changed. Business wants quick answers; they want to test an idea, cross it off if it fails, and move on to the next one. In addition, as IT transitions to the cloud, lack of visibility into the internals of cloud databases and applications make it virtually impossible to implement ETL-based s. Also, the transition to the cloud means greater value is placed on realtime updates, something primarily batch-oriented ETL tools cannot easily deliver. In the digital era, responsiveness is the name of the game, with new requirements arising faster than ever before. There s simply no time to read data from one system, copy it over a network, and write it into a new system. Repetitious, error-prone, timeconsuming, and expensive, ETL tools represent a serious bottleneck. It s not uncommon for IT teams to finish an ETL job only to find it s no longer necessary. With the emergence of application programming interfaces (APIs) and Software as a Service (SaaS), developers no longer have to start from scratch every time they write a program.

WHITE PAPER 5 data and Now they can contract our parts of the work to remote software that can do it better. Tableau Software produces a family of interactive data visualization products focused on business intelligence. Tableau allows you to extract data into Tableau s fast in-memory data engine where you can do ad-hoc visualization at interactive speeds. With this approach, you can query an extract of data without waiting for Hadoop s MapReduce queries to complete. A great strength of Tableau is its ability to connect to data sources directly. In this way, the business users are able to explore the data sources directly and provide value to business quickly. A powerful and popular tool, Tableau works very well as an in-house data-blending tool for smaller data sets. In the realm of truly big data, when data sets become unwieldy or diverse, however, the tool begins to falter. For example, with Tableau data-blending, it s not possible to perform different join operations as well as blend datasources with millions of records on each side. Also, the ability to store historical data, as in the case of tracking changes in data, is strongly limited when using Tableau extracts. At a certain threshold, it no longer makes sense to use Tableau alone. At that point, it becomes necessary to add a logical data warehouse tool such as. A logical data warehouse represents a new data management architecture for analytics which combines the strengths of traditional repository warehouses with alternative data management and access strategies. This new approach is made possible by the maturation of today s networks which are now

WHITE PAPER HOW TO INTEGRATE DATA INTO TABLEAU 6 data and sufficiently fast, reliable, and inter-operable. Logical data warehouse solutions usually involve advanced forms of data, including federation and virtualization, which are key to unifying multiple data ecosystems. The main advantage of the logical data warehouse is that virtual views can be altered without needed to first transform and reload data. View technologies go hand in hand with in-memory functions, and data can be created, processed, and delivered onthe fly, enabling purely semantic views of data structures. Lightly persisted data can be materialized into the view on an asneeded basis. In situations where data is time-sensitive, such as determining production yield on a shop floor, data virtualization techniques can produce results that are several seconds to a few minutes old. With optimization capabilities, execution speeds for queries can be increased ten-fold or more. Another important concept in the logical data warehouse is that of an in-memory data fabric that stretches around the technology stack and around key applications like finance, CRM, ERP, or a call center, for example. The data fabric provides a unified view or collection of views of data in multiple systems across an enterprise, or one look into the big picture. This makes it possible for BI managers to see into multiple databases, applications, and legacy platforms. As the interface layer, it s invisible to the user, who doesn t know if the data is persisted, fetched, materialized, or what not. This offers great scalability, flexibility, and speed for time-sensitive business practices like lean data management. Data Virtuality offers a logical data warehouse approach that allows organizations to keep the tools they currently have, abstract data from multiple sources, and create virtual views through a Web portal. This enables users to quickly query, share,

WHITE PAPER 7 data and and, most importantly, integrate data, whether it resides in flat files, web services, an Oracle database or on a SQL Server. DataVirtuality allows you to take the next step in data blending joining, correlating, and querying massive amounts of live data on the fly. A self-service business intelligence user can join new data and create new insights. Using a traditional ETL approach, a user would first need to have a clear data model in mind. In contrast, DataVirtuality allows you to test new approaches as you think of them, see different angles, and experiment as you go. Here, the direct access possibilities of both Tableau and DataVirtuality create a perfect combination: by using Tableau to access hundreds of data sources directly through DataVirtuality the data exploration possibilities rise a hitherto unknown degree. What really sets DataVirtuality apart, however, is the solution s ability to realize and remember the data being queried. For example, the server can remember that a certain join was used with postgres and Oracle in Tableau. As a result, the user can access that model from internal data storage with a single mouse click, as needed. This kind of immediacy is revolutionary. BI managers no longer have to create or populate models from scratch over and over again. In the past, if you wanted to try new combinations of data, you had to for plan it. You couldn t simply analyze your data immediately. You d need to have your IT department prepare the data for you. With, querying takes mere seconds, with the option to optimize execution speeds for results up to ten times faster than straight querying. However, data exploration is not the only aspect of analytics which benefits from the combination of and Tabeau. Another important aspect is the possibility to define centralized data models to be shared by all Tableau and Non-

WHITE PAPER 8 data and Tableau reports. Finally, businesses can define once and for all how their KPIs are calculated, so that all business users have a single source of truth. Using, a user can build a data model over completely diverse data sources and join the data on the fly. They can pull the data in from the virtual layer, build a new data model using Tableau s own query builder, and define relationships in and among the data sets. This kind of virtual data modeling is not possible for disparate data sources using Tableau alone. If a number of tables, for example, were being ingested from a single system like MySQL, then it would be possible. But when there are a number of different data sources, you need. This is a big data model; it goes further than simple data blending. with big data insights derived from Tableau and As we ve seen, organizations have a historical opportunity to mine big data to transform their business. Data can reveal new business opportunities and dramatically reduce costs. Modern solutions based on the logical data warehouse can give organizations a significant leg up in the race to gain insight and real business advantage from their data. In the past, serious expense and time was needed to upgrade existing BI infrastructures. Today, this is not the case. Datavirtuality automatically builds most of the database structures, creating a virtual layer around all data sources that allows users to start experimenting with the data immediately. The system then observes how the users work with the data and automatically arranges the data structures in the fastest way based on usage patterns. The system also allows users to create

WHITE PAPER 9 data and their own data models and manage them centrally. As an platform, requires a front-end tool such as Excel, Tableau, Looker or QlikView to visualize the data. All common front-end solutions currently on the market can be connected to. The beauty of lies in its flexible nature. It works best when it has direct access to data sources and APIs. Using and Tableau together transforms requirements gathering. What was once a painful, error-prone, and arduous process becomes a data exploration exercise with data profiling tools converging and merging with each other. Users gather requirements by looking at data through a logical layer in a virtual fashion, putting it together in minutes to show the business lead looking over their shoulder. The logical data warehouse enables us to think differently about data and development methods and employ agile development. By enveloping the data of the source system with a virtual view, presents data of all source systems in what appears to be a single large, relational database, which can be handled in a unified way. With user requests translated transparently into queries to the diverse source systems, analyzes user behavior continually and automatically builds an internal data warehouse, ready with answer at the click of a mouse. SOURCES http://www.data.info/data- http://computer.howstuffworks.com/data-4.htm