Introduction to Data Science

Similar documents
Database Vs. Data Warehouse

The Definitive Guide to Preparing Your Data for Tableau

Self-Service Data Preparation for Qlik. Cookbook Series Self-Service Data Preparation for Qlik

Shine a Light on Dark Data with Vertica Flex Tables

SAP Agile Data Preparation Simplify the Way You Shape Data PUBLIC

Modern Data Warehouse The New Approach to Azure BI

Virtuoso Infotech Pvt. Ltd.

Optimized Data Integration for the MSO Market

How to integrate data into Tableau

Fast Innovation requires Fast IT

Xcelerated Business Insights (xbi): Going beyond business intelligence to drive information value

Exploiting and Gaining New Insights for Big Data Analysis

CHAPTER 3 Implementation of Data warehouse in Data Mining

ASG WHITE PAPER DATA INTELLIGENCE. ASG s Enterprise Data Intelligence Solutions: Data Lineage Diving Deeper

Top of Minds Report series Data Warehouse The six levels of integration

TECHNOLOGY BRIEF: CA ERWIN DATA PROFILER. Combining Data Profiling and Data Modeling for Better Data Quality

Drawing the Big Picture

PERSPECTIVE. Data Virtualization A Potential Antidote for Big Data Growing Pains. Abstract

Hybrid Data Platform

Lambda Architecture for Batch and Stream Processing. October 2018

RSDs vs Dossiers Best Practices on When and Where to use them

WKU-MIS-B10 Data Management: Warehousing, Analyzing, Mining, and Visualization. Management Information Systems

UNLEASHING THE VALUE OF THE TERADATA UNIFIED DATA ARCHITECTURE WITH ALTERYX

OLAP Introduction and Overview

What s New in Spotfire DXP 1.1. Spotfire Product Management January 2007

Six Core Data Wrangling Activities. An introductory guide to data wrangling with Trifacta

Embedded Technosolutions

The Emerging Data Lake IT Strategy

ALIGNING CYBERSECURITY AND MISSION PLANNING WITH ADVANCED ANALYTICS AND HUMAN INSIGHT

Top Five Reasons for Data Warehouse Modernization Philip Russom

Data warehouse architecture consists of the following interconnected layers:

LOG FILE ANALYSIS USING HADOOP AND ITS ECOSYSTEMS

Massive Data Analysis

Listening for the Right Signals Using Event Stream Processing for Enterprise Data

Big Data Technology Ecosystem. Mark Burnette Pentaho Director Sales Engineering, Hitachi Vantara

TDWI strives to provide course books that are contentrich and that serve as useful reference documents after a class has ended.

WHAT CIOs NEED TO KNOW TO CAPITALIZE ON HYBRID CLOUD

Empowering Self-Service Capabilities with Agile Analytics

Advanced Solutions of Microsoft SharePoint Server 2013

Unified Governance for Amazon S3 Data Lakes

Question Bank. 4) It is the source of information later delivered to data marts.

Integrating Oracle Databases with NoSQL Databases for Linux on IBM LinuxONE and z System Servers

What s the Value of Your Data? The Agile Advantage

Data Warehouses and Deployment

Database Systems: Design, Implementation, and Management Tenth Edition. Chapter 1 Database Systems

Data Mining. Introduction. Hamid Beigy. Sharif University of Technology. Fall 1395

1 DATAWAREHOUSING QUESTIONS by Mausami Sawarkar

EMC Documentum xdb. High-performance native XML database optimized for storing and querying large volumes of XML content

CTL.SC4x Technology and Systems

Management Information Systems Review Questions. Chapter 6 Foundations of Business Intelligence: Databases and Information Management

Datameer for Data Preparation:

How to Accelerate Merger and Acquisition Synergies

Data Mining. Introduction. Hamid Beigy. Sharif University of Technology. Fall 1394

Data Analyst Nanodegree Syllabus

Value of Data Transformation. Sean Kandel, Co-Founder and CTO of Trifacta

Guide Users along Information Pathways and Surf through the Data

Advanced Solutions of Microsoft SharePoint Server 2013 Course Contact Hours

Using Threat Analytics to Protect Privileged Access and Prevent Breaches

Advanced Solutions of Microsoft SharePoint 2013

The Seven Steps to Implement DataOps

Automate Transform Analyze

Microsoft Developer Day

The #1 Key to Removing the Chaos. in Modern Analytical Environments

Built for Speed: Comparing Panoply and Amazon Redshift Rendering Performance Utilizing Tableau Visualizations

Total Cost of Ownership: Benefits of the OpenText Cloud

Full file at

The Six Principles of BW Data Validation

Big Data - Some Words BIG DATA 8/31/2017. Introduction

EXTRACT DATA IN LARGE DATABASE WITH HADOOP

Data Management Glossary

Data modelling patterns used for integration of operational data stores

Why you should design your data hub top-down vs. bottom-up

Strategic Briefing Paper Big Data

The Business Value of Metadata for Data Governance: The Challenge of Integrating Packaged Applications

Q1) Describe business intelligence system development phases? (6 marks)

Data Warehousing. Ritham Vashisht, Sukhdeep Kaur and Shobti Saini

Introduction to Big-Data

Fraud Mobility: Exploitation Patterns and Insights

Data Analyst Nanodegree Syllabus

Data Mining: Approach Towards The Accuracy Using Teradata!

Solution Brief. Bridging the Infrastructure Gap for Unstructured Data with Object Storage. 89 Fifth Avenue, 7th Floor. New York, NY 10003

DATA VAULT MODELING GUIDE

EMC GREENPLUM MANAGEMENT ENABLED BY AGINITY WORKBENCH

2 The IBM Data Governance Unified Process

Extending the Value of MDM Through Data Virtualization

EUROPEAN ICT PROFESSIONAL ROLE PROFILES VERSION 2 CWA 16458:2018 LOGFILE

Overview of Web Mining Techniques and its Application towards Web

Total Cost of Ownership: Benefits of ECM in the OpenText Cloud

PROJECT PERIODIC REPORT

INTRODUCTORY INFORMATION TECHNOLOGY ENTERPRISE DATABASES AND DATA WAREHOUSES. Faramarz Hendessi

Information Management Fundamentals by Dave Wells

Microsoft SharePoint Server 2013 Plan, Configure & Manage

Inbound Website. How to Build an. Track 1 SEO and SOCIAL

Department of Information Technology, St. Joseph s College (Autonomous), Trichy, TamilNadu, India

Enterprise Application Viewer (eav )

TDWI Data Modeling. Data Analysis and Design for BI and Data Warehousing Systems

Partner Presentation Faster and Smarter Data Warehouses with Oracle OLAP 11g

Is Your Data Viable? Preparing Your Data for SAS Visual Analytics 8.2

Big Data Analytics: What is Big Data? Stony Brook University CSE545, Fall 2016 the inaugural edition

Challenges and Opportunities with Big Data. By: Rohit Ranjan

Transcription:

UNIT I INTRODUCTION TO DATA SCIENCE Syllabus Introduction of Data Science Basic Data Analytics using R R Graphical User Interfaces Data Import and Export Attribute and Data Types Descriptive Statistics Exploratory Data Analysis Visualization Before Analysis Dirty Data Visualizing a Single Variable Examining Multiple Variables Data Exploration Versus Presentation.. Introduction to Data Science Big Data Overview Data is created constantly, and at an ever-increasing rate. Mobile phones, social media, imaging technologies to determine a medical diagnosis all these and more create new data, and that must be stored somewhere for some purpose. Devices and sensors automatically generate diagnostic information that needs to be stored and processed in real time. Merely keeping up with this huge influx of data is difficult, but substantially more challenging is analyzing vast amounts of it, especially when it does not conform to traditional notions of data structure, to identify meaningful patterns and extract useful information. These challenges of the data deluge present the opportunity to transform business, government, science, and everyday life. Several industries have led the way in developing their ability to gather and exploit data: Credit card companies monitor every purchase their customers make and can identify fraudulent purchases with a high degree of accuracy using rules derived by processing billions of transactions. Mobile phone companies analyze subscribers calling patterns to determine, for example, whether a caller s frequent contacts are on a rival network. If that rival network is offering an attractive promotion that might cause the subscriber to defect, the mobile phone company can proactively offer the subscriber an incentive to remain in her contract. For companies such as LinkedIn and Facebook, data itself is their primary product. The valuations of these companies are heavily derived from the data they gather and host, which contains more and more intrinsic value as the data grows. Ms. Selva Mary. G Page 1

Three attributes stand out as defining Big Data characteristics: Huge volume of data (Volume): Rather than thousands\ or millions of rows, Big Data can be billions of rows and millions of columns. Complexity of data types and structures (Variety): Big Data reflects the variety of new data sources, formats, and structures, including digital traces being left on the web and other digital repositories for subsequent analysis. Speed of new data creation and growth (Velocity): Big Data can describe high velocity data, with rapid data ingestion and near real time analysis. Although the volume of Big Data tends to attract the most attention, generally the variety and velocity of the data provide a more apt definition of Big Data. (Big Data is sometimes described as having 3 Vs: volume, variety, and velocity.) Due to its size or structure, Big Data cannot be efficiently analyzed using only traditional databases or methods. Big Data problems require new tools and technologies to store, manage, and realize the business benefit. These new tools and technologies enable creation, manipulation, and management of large datasets and the storage environments that house them. Definition Definition of Big Data comes from the McKinsey Global report from 2011: Big Data is data whose scale, distribution, diversity, and/or timeliness require the use of new technical architectures and analytics to enable insights that unlock new sources of business value. McKinsey s definition of Big Data implies that organizations will need new data architectures and analytic sandboxes, new tools, new analytical methods, and an integration of multiple skills into the new role of the data scientist, Figure 1-1 highlights several sources of the Big Data deluge. Ms. Selva Mary. G Page 2

Different Data Structures Structured data: Data containing a defined data type, format, and structure (that is, transaction data, online analytical processing [OLAP] data cubes, traditional RDBMS, CSV files, and even simple spreadsheets). FIGURE Example of structured data Ms. Selva Mary. G Page 3

Semi-structured data: Textual data files with a discernible pattern that enables parsing (such as Extensible Markup Language [XML] data files that are self-describing and defined by an XML schema). Figure Semi-structured data Ms. Selva Mary. G Page 4

Quasi-structured data: Textual data with erratic data formats that can be formatted with effort, tools, and time (for instance, web clickstream data that may contain inconsistencies in data values and formats). Figure Quasi-structured data Unstructured data: Data that has no inherent structure, which may include text documents, PDFs, images, and video. Figure Unstructured data Ms. Selva Mary. G Page 5

Current Analytical Architecture Data Science projects need workspaces that are purpose-built for experimenting with data, with flexible and agile data architectures. Most organizations still have data warehouses that provide excellent support for traditional reporting and simple data analysis activities but unfortunately have a more difficult time supporting more robust analyses. A typical analytical data architecture that may exist within an organization is explained below Figure shows typical data architecture and several of the challenges it presents to data scientists and others trying to do advanced analytics. This section examines the data flow to the Data Scientist and how this individual fits into the process of getting data to analyze on projects. 1. For data sources to be loaded into the data warehouse, data needs to be well understood, structured, and normalized with the appropriate data type definitions. Although this kind of centralization enables security, backup, and failover of highly critical data, it also means that data typically must go through significant preprocessing and checkpoints before it can enter this sort of controlled environment, which does not lend itself to data exploration and iterative analytics. 2. As a result of this level of control on the EDW, additional local systems may emerge in the form of departmental warehouses and local data marts that business users create to accommodate their need for flexible analysis. These local data marts may not have the same constraints for security and structure as the main EDW and allow users to do some level of more in-depth analysis. However, these one-off systems reside in isolation, often are not synchronized or integrated with other data stores, and may not be backed up. Ms. Selva Mary. G Page 6

3. Once in the data warehouse, data is read by additional applications across the enterprise for BI and reporting purposes. These are high-priority operational processes getting critical data feeds from the data warehouses and repositories. 4. At the end of this workflow, analysts get data provisioned for their downstream analytics. Because users generally are not allowed to run custom or intensive analytics on production databases, analysts create data extracts from the EDW to analyze data offline in R or other local analytical tools. Many times these tools are limited to in-memory analytics on desktops analyzing samples of data, rather than the entire population of a dataset. Because these analyses are based on data extracts, they reside in a separate location, and the results of the analysis and any insights on the quality of the data or anomalies rarely are fed back into the main data repository. Role of Data Scientists Data scientists play the most active roles in the four A s of data: Data architecture A data scientist would help the system architect by providing input on how the data would need to be routed and organized to support the analysis, visualization, and presentation of the data to the appropriate people. Data acquisition Representing, transforming, grouping, and linking the data are all tasks that need to occur before the data can be profitably analyzed, and these are all tasks in which the data scientist is actively involved. Data analysis The analysis phase is where data scientists are most heavily involved. In this context we are using analysis to include summarization of the data, using portions of data (samples) to make inferences about the larger context, and visualization of the data by presenting it in tables, graphs, and even animations. Data archiving Finally, the data scientist must become involved in the archiving of the data. Preservation of collected data in a form that makes it highly reusable - what you might think of as "data Ms. Selva Mary. G Page 7

duration"- is a difficult challenge because it is so hard to anticipate all of the future uses of the data. The data science process The data science process typically consists of six steps, as you can see in the mind map in figure 1. Setting the research goal Data science is mostly done in the context of an organization. When the business asks you to perform a data science project, you ll first prepare a project charter. This charter contains information such as what you re going to research, how the company benefits from that, what data and resources you need, a timetable, and deliverables. 2. Retrieving data The second step is to collect data. You ve stated in the project charter which data you need and where you can find it. In this step you ensure that you can use the data in your program, which Ms. Selva Mary. G Page 8

means checking the existence, quality, and access to the data. Data can also be delivered by thirdparty companies and take many forms ranging from Excel spreadsheets to different types of databases. 3. Data cleansing Data collection is an error-prone process; in this phase you enhance the quality of the data and prepare it for use in subsequent steps. This phase consists of three subphases: data cleansing removes false values from a data source and inconsistencies across data sources, data integration enriches data sources by combining information from multiple data sources, and data transformation ensures that the data is in a suitable format for use in your models. 4. Data exploration Data exploration is concerned with building a deeper understanding of your data. You try to understand how variables interact with each other, the distribution of the data, and whether there are outliers. To achieve this you mainly use descriptive statistics, visual techniques, and simple modeling. This step often goes under the abbreviation EDA for Exploratory Data Analysis. 5. Data modeling or model building In this phase you use models, domain knowledge, and insights about the data you found in the previous steps to answer the research question. You select a technique from the fields of statistics, machine learning, operations research, and so on. Building a model is an iterative step between selecting the variables for the model, executing the model, and model diagnostics. 6. Presentation and automation Finally, you present the results to your business. These results can take many forms, ranging from presentations to research reports. Sometimes you ll need to automate the execution of the process because the business will want to use the insights you gained in another project or enable an operational process to use the outcome from your model. AN ITERATIVE PROCESS The previous description of the data science process gives you the impression that you walk through this process in a linear way, but in reality you often have to step back and rework certain findings. For instance, you might find outliers in the data exploration phase that point to data import errors. As part of the data science process you gain incremental insights, which may lead to new questions. To prevent rework, make sure that you scope the business question clearly and thoroughly at the start. Ms. Selva Mary. G Page 9

Challenges in Data Science Preparing Data (Noisy, Incomplete, Diverse, Streaming ) Analyze Data (Scalable, Accurate, Realtime, Advanced Methods, Probabilities & Uncertainties) Represent Analysis Results (i.e. data product) (Story-telling, Interactive, explainable ) **************************** Ms. Selva Mary. G Page 10