RECORD LINKAGE, A REAL USE CASE WITH SPARK ML. Alexis Seigneurin
|
|
- Malcolm Mills
- 6 years ago
- Views:
Transcription
1 RECORD LINKAGE, A REAL USE CASE WITH SPARK ML Alexis Seigneurin
2 Who I am Software engineer for 15 years Consultant at Ippon USA, previously at Ippon France Favorite subjects: Spark, Machine Learning, Cassandra Spark
3 200 software engineers in France and the US In the US: offices in DC, NYC and Richmond, Virginia Digital, Big Data and Cloud applications Java & Agile expertise Open-source projects: JHipster, Tatami,
4 The project Record Linkage Record Linkage with Machine learning Use cases: Entity resolution Deduplication Entity disambiguation Find new clients who come from insurance comparison services Commission Find duplicates in existing files (acquisitions)
5 Overview
6 Purpose Find duplicates! ID veh codptgar_veh dt_nais_cp dt_permis_cp redmaj_cp formule PE TIERS... FO VOL_INCENDIE... FI TOUS_RISQUES... RE TIERS... FO TOUS_RISQUES
7 Steps 1. Preprocessing 1. Find potential duplicates 2. Feature engineering 2. Manual labeling of a sample 3. Machine Learning to make predictions on the rest of the records
8 Prototype Crafted by a Data Scientist Not architectured, not versioned, not unit tested Not ready for production Spark, but a lot of Spark SQL (data processing) Machine Learning in Python (Scikit Learn) Objective: industrialization of the code
9 Preprocessing
10 Inputs Data (CSV) + Schema (JSON) ;Jose;Lester;10/10/ ;José;Lester;10/10/ ;Tyler;Hunt;12/12/ ;Tiler;Hunt;25/12/ ;Patrick;Andrews; { "tableschemabeforeselection": [ { "name": "ID", "typefield": "StringType", "hardjoin": false }, { "name": "name", "typefield": "StringType", "hardjoin": true, "cleaning": "digitletter", "listfeature": [ "scarcity" ], "listdistance": [ "equality", "soundlike" ] },...
11 Data loading Spark CSV module DataFrame Don t use type inference ID name surname birthdt Jose Lester 10/10/ José Lester 10/10/ Tyler Hunt 12/12/ Tiler Hunt 25/12/ Patrick Andrews
12 Data cleansing Parsing of dates, numbers Cleaning of strings ID name surname birthdt jose lester jose lester tyler hunt tiler hunt patrick andrews null
13 Feature calculation Convert strings to phonetics (Beider-Morse) ID name surname birthdt BMencoded_name jose lester ios iosi ioz iozi jose lester ios iosi ioz iozi tyler hunt tilir tiler hunt tqlir tili tilir patrick andrews null pytrqk pytrik pat
14 Find potential duplicates Auto-join (more on that later ) ID_1 name_1 surname_1... ID_2 name_2 surname_ jose lester jose lester tyler hunt tiler hunt
15 Distance calculation Several distance algorithms: Levenshtein distance, date difference ID_1... ID_2... equality_name soundlike_name... datediff_birthdt
16 Standardization / vectorization Standardization of distances only Vectorization (2 vectors) ID_1 name_1 surname_1 birthdt_1 ID_2 name_2 surname_2 birthdt_2 distances other_features jose lester jose lester [0.0,0.0,... [2.0,2.0, tyler hunt tiler hunt [1.0,1.0,... [1.0,2.0,
17 Spark SQL DataFrames
18 From SQL Generated SQL requests Hard to maintain (especially as regards to UDFs) val cleaningrequest = tableschema.map(x => { x.cleaningfuction match { case (Some(a), _) => a + "(" + x.name + ") as " + x.name case _ => x.name } }).mkstring(", ") val cleanedtable = sqlcontext.sql("select " + cleaningrequest + " from " + tablename) cleanedtable.registertemptable(schema.tablename + "_cleaned")
19 to DataFrames DataFrame primitives More work done by the Scala compiler val cleaneddf = tableschema.filter(_.cleaning.isdefined).foldleft(df) { case (df, field) => val udf: UserDefinedFunction =... // get the cleaning UDF } df.withcolumn(field.name + "_cleaned", udf.apply(df(field.name))).drop(field.name).withcolumnrenamed(field.name + "_cleaned", field.name)
20 Unit testing
21 Unit testing Scalatest + Scoverage Coverage of all the data processing operations
22 Comparison of Row objects ;Jose;Lester;10/10/ ;Jose =-+;Lester éà;10/10/ ;Jose;Lester;invalid date val resdf = schema.cleantable(rows) "The cleaning process" should "clean text fields" in { val res = resdf.select("id", "name", "surname").collect() val expected = Array( Row("000010", "jose", "lester"), Row("000011", "jose", "lester ea"), Row("000012", "jose", "lester") ) res should contain thesameelementsas expected } "The cleaning process" should "parse dates" in {...
23 Shared SparkContext Don ts: Use one SparkContext per class of tests multiple contexts Setup / tear down the SparkContext for each test slow tests Do s: Use a shared SparkContext object SparkTestContext { val conf = new SparkConf().setAppName("deduplication-tests").setMaster("local[*]") val sc = new SparkContext(conf) val sqlcontext = new SQLContext(sc) }
24 spark-testing-base Holden Karau s spark-testing-base library Provides: Shared SparkContext and SQLContext Comparison of RDDs, DataFrames, DataSets Mock data generators
25 Matching potential duplicates
26 Join strategy Prospects New clients For record linkage, first merge the two sources Then auto-join Duplicate
27 Join - Volume of data 100 Input: 1M records Cartesian product: 1000 B records Find an appropriate join condition 25 0
28 Join condition Multiples join on 2 fields Equality of values or custom condition (UDF) Union between all the intermediate results E.g. with fields name, surname, birth_date: df1.join(df2, (df1("id_1") < df2("id_2")) && (df1("name_1") === df2("name_2")) && (soundlike(df1("surname_1"), df2("surname_2"))) df1.join(df2, (df1("id_1") < df2("id_2")) && (df1("name_1") === df2("name_2")) && (df1("birth_date_1") === df2("birth_date_2"))) UNION df1.join(df2, (df1("id_1") < df2("id_2")) && (soundlike(df1("surname_1"), df2("surname_2"))) && (df1("birth_date_1") === df2("birth_date_2")))
29 DataFrames extension
30 DataFrames extension 3 types of columns Data Non-distance features Distances ID_1... ID_2... equality_name soundlike_name... datediff_birthdt
31 DataFrames extension DataFrame columns have a name and a data type DataFrameExt = DataFrame + metadata over columns case class OutputColumn(name: String, columntype: ColumnType) class DataFrameExt(val df: DataFrame, val outputcolumns: Seq[OutputColumn]) { def show() = df.show() def drop(colname: String): DataFrameExt =... def withcolumn(colname: String, col: Column, columntype: ColumnType): DataFrameExt =......
32 Labeling
33 Labeling Manual operation Is this a duplicate? Yes / No Performed on a sample of the potential duplicates Between 1000 and records
34 Labeling
35 Predictions
36 Predictions Machine Learning Random Forests (Gradient Boosting Trees also give good results) Training on the potential duplicates labeled by hand Predictions on the potential duplicates not labeled by hand
37 Predictions Sample: 1000 records Training set: 800 records Test set: 200 records Results Found 53 duplicates on the 58 True positives: 53 expected (53+5) and only 2 errors False positives: 2 True negatives: 126 False negatives: 5 Precision 93% Recall 91%
38 Summary & Conclusion
39 Summary Single engine for Record Linkage and Deduplication Machine Learning Specific rules for each dataset Higher identification of matches Previously ~50% Now ~90%
40 Thank
THE CONTRACTOR IS ACTING UNDER A FRAMEWORK CONTRACT CONCLUDED WITH THE COMMISSION
Apache Spark Lorenzo Di Gaetano THE CONTRACTOR IS ACTING UNDER A FRAMEWORK CONTRACT CONCLUDED WITH THE COMMISSION What is Apache Spark? A general purpose framework for big data processing It interfaces
More informationSpark 2. Alexey Zinovyev, Java/BigData Trainer in EPAM
Spark 2 Alexey Zinovyev, Java/BigData Trainer in EPAM With IT since 2007 With Java since 2009 With Hadoop since 2012 With EPAM since 2015 About Secret Word from EPAM itsubbotnik Big Data Training 3 Contacts
More informationspark-testing-java Documentation Release latest
spark-testing-java Documentation Release latest Nov 20, 2018 Contents 1 Input data preparation 3 2 Java 5 2.1 Context creation............................................. 5 2.2 Data preparation.............................................
More informationWHAT S NEW IN SPARK 2.0: STRUCTURED STREAMING AND DATASETS
WHAT S NEW IN SPARK 2.0: STRUCTURED STREAMING AND DATASETS Andrew Ray StampedeCon 2016 Silicon Valley Data Science is a boutique consulting firm focused on transforming your business through data science
More information732A54 Big Data Analytics: SparkSQL. Version: Dec 8, 2016
732A54 Big Data Analytics: SparkSQL Version: Dec 8, 2016 2016-12-08 2 DataFrames A DataFrame is a distributed collection of data organized into named columns. It is conceptually equivalent to a table in
More informationAgenda. Spark Platform Spark Core Spark Extensions Using Apache Spark
Agenda Spark Platform Spark Core Spark Extensions Using Apache Spark About me Vitalii Bondarenko Data Platform Competency Manager Eleks www.eleks.com 20 years in software development 9+ years of developing
More informationSpark and Spark SQL. Amir H. Payberah. SICS Swedish ICT. Amir H. Payberah (SICS) Spark and Spark SQL June 29, / 71
Spark and Spark SQL Amir H. Payberah amir@sics.se SICS Swedish ICT Amir H. Payberah (SICS) Spark and Spark SQL June 29, 2016 1 / 71 What is Big Data? Amir H. Payberah (SICS) Spark and Spark SQL June 29,
More informationAsanka Padmakumara. ETL 2.0: Data Engineering with Azure Databricks
Asanka Padmakumara ETL 2.0: Data Engineering with Azure Databricks Who am I? Asanka Padmakumara Business Intelligence Consultant, More than 8 years in BI and Data Warehousing A regular speaker in data
More informationHigher level data processing in Apache Spark
Higher level data processing in Apache Spark Pelle Jakovits 12 October, 2016, Tartu Outline Recall Apache Spark Spark DataFrames Introduction Creating and storing DataFrames DataFrame API functions SQL
More information2/26/2017. DataFrame. The input data can be queried by using
Spark SQL is the Spark component for structured data processing It provides a programming abstraction called s and can act as distributed SQL query engine The input data can be queried by using Ad-hoc
More informationSparkling Water. August 2015: First Edition
Sparkling Water Michal Malohlava Alex Tellez Jessica Lanford http://h2o.gitbooks.io/sparkling-water-and-h2o/ August 2015: First Edition Sparkling Water by Michal Malohlava, Alex Tellez & Jessica Lanford
More informationAnalytic Cloud with. Shelly Garion. IBM Research -- Haifa IBM Corporation
Analytic Cloud with Shelly Garion IBM Research -- Haifa 2014 IBM Corporation Why Spark? Apache Spark is a fast and general open-source cluster computing engine for big data processing Speed: Spark is capable
More informationParallel Processing Spark and Spark SQL
Parallel Processing Spark and Spark SQL Amir H. Payberah amir@sics.se KTH Royal Institute of Technology Amir H. Payberah (KTH) Spark and Spark SQL 2016/09/16 1 / 82 Motivation (1/4) Most current cluster
More information@h2oai presents. Sparkling Water Meetup
@h2oai & @mmalohlava presents Sparkling Water Meetup User-friendly API for data transformation Large and active community Memory efficient Performance of computation Platform components - SQL Machine learning
More informationProcessing of big data with Apache Spark
Processing of big data with Apache Spark JavaSkop 18 Aleksandar Donevski AGENDA What is Apache Spark? Spark vs Hadoop MapReduce Application Requirements Example Architecture Application Challenges 2 WHAT
More informationIntroduction to Spark
Introduction to Spark Outlines A brief history of Spark Programming with RDDs Transformations Actions A brief history Limitations of MapReduce MapReduce use cases showed two major limitations: Difficulty
More informationPyspark standalone code
COSC 6339 Big Data Analytics Introduction to Spark (II) Edgar Gabriel Spring 2017 Pyspark standalone code from pyspark import SparkConf, SparkContext from operator import add conf = SparkConf() conf.setappname(
More informationDatabases and Big Data Today. CS634 Class 22
Databases and Big Data Today CS634 Class 22 Current types of Databases SQL using relational tables: still very important! NoSQL, i.e., not using relational tables: term NoSQL popular since about 2007.
More informationIBM Data Science Experience White paper. SparkR. Transforming R into a tool for big data analytics
IBM Data Science Experience White paper R Transforming R into a tool for big data analytics 2 R Executive summary This white paper introduces R, a package for the R statistical programming language that
More informationApache Ignite - Using a Memory Grid for Heterogeneous Computation Frameworks A Use Case Guided Explanation. Chris Herrera Hashmap
Apache Ignite - Using a Memory Grid for Heterogeneous Computation Frameworks A Use Case Guided Explanation Chris Herrera Hashmap Topics Who - Key Hashmap Team Members The Use Case - Our Need for a Memory
More informationOverview. Prerequisites. Course Outline. Course Outline :: Apache Spark Development::
Title Duration : Apache Spark Development : 4 days Overview Spark is a fast and general cluster computing system for Big Data. It provides high-level APIs in Scala, Java, Python, and R, and an optimized
More informationIntro To Spark. John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center. Copyright 2017
Intro To Spark John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2017 Performance First, use RAM Also, be smarter Spark Capabilities (i.e. Hadoop shortcomings) Ease of
More informationBlended Learning Outline: Developer Training for Apache Spark and Hadoop (180404a)
Blended Learning Outline: Developer Training for Apache Spark and Hadoop (180404a) Cloudera s Developer Training for Apache Spark and Hadoop delivers the key concepts and expertise need to develop high-performance
More informationMLeap: Release Spark ML Pipelines
MLeap: Release Spark ML Pipelines Hollin Wilkins & Mikhail Semeniuk SATURDAY Web Dev @ Cornell Studied some General Biology Rails Consulting for TrueCar and other companies Implement ML model for ClearBook
More informationUnifying Big Data Workloads in Apache Spark
Unifying Big Data Workloads in Apache Spark Hossein Falaki @mhfalaki Outline What s Apache Spark Why Unification Evolution of Unification Apache Spark + Databricks Q & A What s Apache Spark What is Apache
More informationMLI - An API for Distributed Machine Learning. Sarang Dev
MLI - An API for Distributed Machine Learning Sarang Dev MLI - API Simplify the development of high-performance, scalable, distributed algorithms. Targets common ML problems related to data loading, feature
More informationMatrix Computations and " Neural Networks in Spark
Matrix Computations and " Neural Networks in Spark Reza Zadeh Paper: http://arxiv.org/abs/1509.02256 Joint work with many folks on paper. @Reza_Zadeh http://reza-zadeh.com Training Neural Networks Datasets
More informationIntro To Spark. John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center. Copyright 2017
Intro To Spark John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2017 Spark Capabilities (i.e. Hadoop shortcomings) Performance First, use RAM Also, be smarter Ease of
More informationDatabricks, an Introduction
Databricks, an Introduction Chuck Connell, Insight Digital Innovation Insight Presentation Speaker Bio Senior Data Architect at Insight Digital Innovation Focus on Azure big data services HDInsight/Hadoop,
More informationUsing Apache Phoenix to store and access data
3 Date of Publish: 2018-08-30 http://docs.hortonworks.com Contents ii Contents What's New in Apache Phoenix...4 Orchestrating SQL and APIs with Apache Phoenix...4 Enable Phoenix and interdependent components...4
More informationDATA SCIENCE USING SPARK: AN INTRODUCTION
DATA SCIENCE USING SPARK: AN INTRODUCTION TOPICS COVERED Introduction to Spark Getting Started with Spark Programming in Spark Data Science with Spark What next? 2 DATA SCIENCE PROCESS Exploratory Data
More informationCSE 444: Database Internals. Lecture 23 Spark
CSE 444: Database Internals Lecture 23 Spark References Spark is an open source system from Berkeley Resilient Distributed Datasets: A Fault-Tolerant Abstraction for In-Memory Cluster Computing. Matei
More informationNew Developments in Spark
New Developments in Spark And Rethinking APIs for Big Data Matei Zaharia and many others What is Spark? Unified computing engine for big data apps > Batch, streaming and interactive Collection of high-level
More informationWHITE PAPER. Apache Spark: RDD, DataFrame and Dataset. API comparison and Performance Benchmark in Spark 2.1 and Spark 1.6.3
WHITE PAPER Apache Spark: RDD, DataFrame and Dataset API comparison and Performance Benchmark in Spark 2.1 and Spark 1.6.3 Prepared by: Eyal Edelman, Big Data Practice Lead Michael Birch, Big Data and
More informationA Tutorial on Apache Spark
A Tutorial on Apache Spark A Practical Perspective By Harold Mitchell The Goal Learning Outcomes The Goal Learning Outcomes NOTE: The setup, installation, and examples assume Windows user Learn the following:
More information빅데이터기술개요 2016/8/20 ~ 9/3. 윤형기
빅데이터기술개요 2016/8/20 ~ 9/3 윤형기 (hky@openwith.net) D4 http://www.openwith.net 2 Hive http://www.openwith.net 3 What is Hive? 개념 a data warehouse infrastructure tool to process structured data in Hadoop. Hadoop
More informationStructured Streaming. Big Data Analysis with Scala and Spark Heather Miller
Structured Streaming Big Data Analysis with Scala and Spark Heather Miller Why Structured Streaming? DStreams were nice, but in the last session, aggregation operations like a simple word count quickly
More informationBig Data Analytics with Apache Spark. Nastaran Fatemi
Big Data Analytics with Apache Spark Nastaran Fatemi Apache Spark Throughout this part of the course we will use the Apache Spark framework for distributed data-parallel programming. Spark implements a
More informationDistributed Machine Learning" on Spark
Distributed Machine Learning" on Spark Reza Zadeh @Reza_Zadeh http://reza-zadeh.com Outline Data flow vs. traditional network programming Spark computing engine Optimization Example Matrix Computations
More informationUsing MySQL, Hadoop and Spark for Data Analysis
Using MySQL, Hadoop and Spark for Data Analysis Alexander Rubin Principle Architect, Percona September 21, 2015 About Me Alexander Rubin, Principal Consultant, Percona Working with MySQL for over 10 years
More informationSpark Streaming. Big Data Analysis with Scala and Spark Heather Miller
Spark Streaming Big Data Analysis with Scala and Spark Heather Miller Where Spark Streaming fits in (1) Spark is focused on batching Processing large, already-collected batches of data. For example: Where
More informationChapter 1 - The Spark Machine Learning Library
Chapter 1 - The Spark Machine Learning Library Objectives Key objectives of this chapter: The Spark Machine Learning Library (MLlib) MLlib dense and sparse vectors and matrices Types of distributed matrices
More informationApache Spark is a fast and general-purpose engine for large-scale data processing Spark aims at achieving the following goals in the Big data context
1 Apache Spark is a fast and general-purpose engine for large-scale data processing Spark aims at achieving the following goals in the Big data context Generality: diverse workloads, operators, job sizes
More informationTurning Relational Database Tables into Spark Data Sources
Turning Relational Database Tables into Spark Data Sources Kuassi Mensah Jean de Lavarene Director Product Mgmt Director Development Server Technologies October 04, 2017 3 Safe Harbor Statement The following
More informationCloud Computing 3. CSCI 4850/5850 High-Performance Computing Spring 2018
Cloud Computing 3 CSCI 4850/5850 High-Performance Computing Spring 2018 Tae-Hyuk (Ted) Ahn Department of Computer Science Program of Bioinformatics and Computational Biology Saint Louis University Learning
More informationCertification Study Guide. MapR Certified Spark Developer v1 Study Guide
Certification Study Guide MapR Certified Spark Developer v1 Study Guide 1 CONTENTS About MapR Study Guides... 3 MapR Certified Spark Developer (MCSD)... 3 SECTION 1 WHAT S ON THE EXAM?... 5 1. Load and
More informationDistributed Computing with Spark and MapReduce
Distributed Computing with Spark and MapReduce Reza Zadeh @Reza_Zadeh http://reza-zadeh.com Traditional Network Programming Message-passing between nodes (e.g. MPI) Very difficult to do at scale:» How
More informationBig Data Analytics with Hadoop and Spark at OSC
Big Data Analytics with Hadoop and Spark at OSC 09/28/2017 SUG Shameema Oottikkal Data Application Engineer Ohio SuperComputer Center email:soottikkal@osc.edu 1 Data Analytics at OSC Introduction: Data
More informationData Science with PostgreSQL
Balázs Bárány Data Scientist pgconf.de 2015 Contents Introduction What is Data Science? Process model Tools and methods of Data Scientists Business & data understanding Preprocessing Modeling Evaluation
More informationCSC 261/461 Database Systems Lecture 24. Spring 2017 MW 3:25 pm 4:40 pm January 18 May 3 Dewey 1101
CSC 261/461 Database Systems Lecture 24 Spring 2017 MW 3:25 pm 4:40 pm January 18 May 3 Dewey 1101 Announcements Term Paper due on April 20 April 23 Project 1 Milestone 4 is out Due on 05/03 But I would
More informationHadoop course content
course content COURSE DETAILS 1. In-detail explanation on the concepts of HDFS & MapReduce frameworks 2. What is 2.X Architecture & How to set up Cluster 3. How to write complex MapReduce Programs 4. In-detail
More informationIndex. bfs() function, 225 Big data characteristics, 2 variety, 3 velocity, 3 veracity, 3 volume, 2 Breadth-first search algorithm, 220, 225
Index A Anonymous function, 66 Apache Hadoop, 1 Apache HBase, 42 44 Apache Hive, 6 7, 230 Apache Kafka, 8, 178 Apache License, 7 Apache Mahout, 5 Apache Mesos, 38 42 Apache Pig, 7 Apache Spark, 9 Apache
More informationAn Introduction to Apache Spark
An Introduction to Apache Spark Anastasios Skarlatidis @anskarl Software Engineer/Researcher IIT, NCSR "Demokritos" Outline Part I: Getting to know Spark Part II: Basic programming Part III: Spark under
More informationApache Spark 2.0. Matei
Apache Spark 2.0 Matei Zaharia @matei_zaharia What is Apache Spark? Open source data processing engine for clusters Generalizes MapReduce model Rich set of APIs and libraries In Scala, Java, Python and
More informationReactive App using Actor model & Apache Spark. Rahul Kumar Software
Reactive App using Actor model & Apache Spark Rahul Kumar Software Developer @rahul_kumar_aws About Sigmoid We build realtime & big data systems. OUR CUSTOMERS Agenda Big Data - Intro Distributed Application
More informationBig Data analytics in insurance
Big Data analytics in insurance Who we are Experts At Your Service > Over 50 specialists in IT infrastructure > Certified, experienced, passionate Based In Switzerland > 100% self-financed Swiss company
More informationHADOOP COURSE CONTENT (HADOOP-1.X, 2.X & 3.X) (Development, Administration & REAL TIME Projects Implementation)
HADOOP COURSE CONTENT (HADOOP-1.X, 2.X & 3.X) (Development, Administration & REAL TIME Projects Implementation) Introduction to BIGDATA and HADOOP What is Big Data? What is Hadoop? Relation between Big
More informationCompile-Time Code Generation for Embedded Data-Intensive Query Languages
Compile-Time Code Generation for Embedded Data-Intensive Query Languages Leonidas Fegaras University of Texas at Arlington http://lambda.uta.edu/ Outline Emerging DISC (Data-Intensive Scalable Computing)
More informationStarting with Apache Spark, Best Practices and Learning from the Field
Starting with Apache Spark, Best Practices and Learning from the Field Felix Cheung, Principal Engineer + Spark Committer Spark@Microsoft Best Practices Enterprise Solutions Resilient - Fault tolerant
More informationResearch challenges in data-intensive computing The Stratosphere Project Apache Flink
Research challenges in data-intensive computing The Stratosphere Project Apache Flink Seif Haridi KTH/SICS haridi@kth.se e2e-clouds.org Presented by: Seif Haridi May 2014 Research Areas Data-intensive
More informationMigrate from Netezza Workload Migration
Migrate from Netezza Automated Big Data Open Netezza Source Workload Migration CASE SOLUTION STUDY BRIEF Automated Netezza Workload Migration To achieve greater scalability and tighter integration with
More informationDelving Deep into Hadoop Course Contents Introduction to Hadoop and Architecture
Delving Deep into Hadoop Course Contents Introduction to Hadoop and Architecture Hadoop 1.0 Architecture Introduction to Hadoop & Big Data Hadoop Evolution Hadoop Architecture Networking Concepts Use cases
More informationApache Spark Tutorial
Apache Spark Tutorial Reynold Xin @rxin BOSS workshop at VLDB 2017 Apache Spark The most popular and de-facto framework for big data (science) APIs in SQL, R, Python, Scala, Java Support for SQL, ETL,
More informationOptimizing Across Relational and Linear Algebra in Parallel Analytics Pipelines
Optimizing Across Relational and Linear Algebra in Parallel Analytics Pipelines Asterios Katsifodimos Assistant Professor @ TU Delft a.katsifodimos@tudelft.nl http://asterios.katsifodimos.com Outline Introduction
More informationACHIEVEMENTS FROM TRAINING
LEARN WELL TECHNOCRAFT DATA SCIENCE/ MACHINE LEARNING SYLLABUS 8TH YEAR OF ACCOMPLISHMENTS AUTHORIZED GLOBAL CERTIFICATION CENTER FOR MICROSOFT, ORACLE, IBM, AWS AND MANY MORE. 8411002339/7709292162 WWW.DW-LEARNWELL.COM
More informationAn Introduction to Apache Spark
An Introduction to Apache Spark Amir H. Payberah amir@sics.se SICS Swedish ICT Amir H. Payberah (SICS) Apache Spark Feb. 2, 2016 1 / 67 Big Data small data big data Amir H. Payberah (SICS) Apache Spark
More informationThe input data can be queried by using
1 Spark SQL is the Spark component for structured data processing It provides a programming abstraction called Dataset and can act as a distributed SQL query engine The input data can be queried by using
More informationMachine Learning With Spark
Ons Dridi R&D Engineer 13 Novembre 2015 Centre d Excellence en Technologies de l Information et de la Communication CETIC Presentation - An applied research centre in the field of ICT - The knowledge developed
More informationAdding Native SQL Support to Spark with C talyst. Michael Armbrust
Adding Native SQL Support to Spark with C talyst Michael Armbrust Overview Catalyst is an optimizer framework for manipulating trees of relational operators. Catalyst enables native support for executing
More informationLightning Fast Cluster Computing. Michael Armbrust Reflections Projections 2015 Michast
Lightning Fast Cluster Computing Michael Armbrust - @michaelarmbrust Reflections Projections 2015 Michast What is Apache? 2 What is Apache? Fast and general computing engine for clusters created by students
More information10/05/2018. The following slides show how to
The following slides show how to Create a classification model based on the logistic regression algorithm for textual documents Apply the model to new textual documents The input training dataset represents
More informationCreating a Recommender System. An Elasticsearch & Apache Spark approach
Creating a Recommender System An Elasticsearch & Apache Spark approach My Profile SKILLS Álvaro Santos Andrés Big Data & Analytics Solution Architect in Ericsson with more than 12 years of experience focused
More informationBig data systems 12/8/17
Big data systems 12/8/17 Today Basic architecture Two levels of scheduling Spark overview Basic architecture Cluster Manager Cluster Cluster Manager 64GB RAM 32 cores 64GB RAM 32 cores 64GB RAM 32 cores
More informationData Warehousing. Jens Teubner, TU Dortmund Summer Jens Teubner Data Warehousing Summer
Jens Teubner Data Warehousing Summer 2018 1 Data Warehousing Jens Teubner, TU Dortmund jens.teubner@cs.tu-dortmund.de Summer 2018 Jens Teubner Data Warehousing Summer 2018 160 Part VI ETL Process ETL Overview
More informationStream Processing Platforms Storm, Spark,.. Batch Processing Platforms MapReduce, SparkSQL, BigQuery, Hive, Cypher,...
Data Ingestion ETL, Distcp, Kafka, OpenRefine, Query & Exploration SQL, Search, Cypher, Stream Processing Platforms Storm, Spark,.. Batch Processing Platforms MapReduce, SparkSQL, BigQuery, Hive, Cypher,...
More informationINTERMEDIATE SQL GOING BEYOND THE SELECT. Created by Brian Duffey
INTERMEDIATE SQL GOING BEYOND THE SELECT Created by Brian Duffey WHO I AM Brian Duffey 3 years consultant at michaels, ross, and cole 9+ years SQL user What have I used SQL for? ROADMAP Introduction 1.
More informationEnd-to-End data mining feature integration, transformation and selection with Datameer Datameer, Inc. All rights reserved.
End-to-End data mining feature integration, transformation and selection with Datameer Fastest time to Insights Rapid Data Integration Zero coding data integration Wizard-led data integration & No ETL
More informationStream Processing Platforms Storm, Spark,.. Batch Processing Platforms MapReduce, SparkSQL, BigQuery, Hive, Cypher,...
Data Ingestion ETL, Distcp, Kafka, OpenRefine, Query & Exploration SQL, Search, Cypher, Stream Processing Platforms Storm, Spark,.. Batch Processing Platforms MapReduce, SparkSQL, BigQuery, Hive, Cypher,...
More informationIntroduction to Apache Spark
Introduction to Apache Spark Bu eğitim sunumları İstanbul Kalkınma Ajansı nın 2016 yılı Yenilikçi ve Yaratıcı İstanbul Mali Destek Programı kapsamında yürütülmekte olan TR10/16/YNY/0036 no lu İstanbul
More informationSimplifying ML Workflows with Apache Beam & TensorFlow Extended
Simplifying ML Workflows with Apache Beam & TensorFlow Extended Tyler Akidau @takidau Software Engineer at Google Apache Beam PMC Apache Beam Portable data-processing pipelines Example pipelines Python
More informationCS 696 Intro to Big Data: Tools and Methods Fall Semester, 2016 Doc 25 Spark 2 Nov 29, 2016
CS 696 Intro to Big Data: Tools and Methods Fall Semester, 2016 Doc 25 Spark 2 Nov 29, 2016 Copyright, All rights reserved. 2016 SDSU & Roger Whitney, 5500 Campanile Drive, San Diego, CA 92182-7700 USA.
More informationUsing Spark SQL in Spark 2
201704 Using Spark SQL in Spark 2 Spark SQL in Spark 2: Overview and SparkSession Chapter 1 Course Chapters Spark SQL in Spark 2: Overview and SparkSession Spark SQL in Spark 2: Datasets, DataFrames, and
More informationKafka pours and Spark resolves! Alexey Zinovyev, Java/BigData Trainer in EPAM
Kafka pours and Spark resolves! Alexey Zinovyev, Java/BigData Trainer in EPAM With IT since 2007 With Java since 2009 With Hadoop since 2012 With Spark since 2014 With EPAM since 2015 About Contacts E-mail
More informationRapid growth of massive datasets
Overview Rapid growth of massive datasets E.g., Online activity, Science, Sensor networks Data Distributed Clusters are Pervasive Data Distributed Computing Mature Methods for Common Problems e.g., classification,
More informationCERTIFICATE IN SOFTWARE DEVELOPMENT LIFE CYCLE IN BIG DATA AND BUSINESS INTELLIGENCE (SDLC-BD & BI)
CERTIFICATE IN SOFTWARE DEVELOPMENT LIFE CYCLE IN BIG DATA AND BUSINESS INTELLIGENCE (SDLC-BD & BI) The Certificate in Software Development Life Cycle in BIGDATA, Business Intelligence and Tableau program
More informationStatic type safety guarantees for the operators of a relational database querying system. Cédric Lavanchy
Static type safety guarantees for the operators of a relational database querying system Cédric Lavanchy June 6, 2008 Contents 1 Previous work 2 2 Goal 3 3 Theory bases 4 3.1 Typing a relation...........................
More informationOpen Data Standards for Administrative Data Processing
University of Pennsylvania ScholarlyCommons 2018 ADRF Network Research Conference Presentations ADRF Network Research Conference Presentations 11-2018 Open Data Standards for Administrative Data Processing
More informationData Science Bootcamp Curriculum. NYC Data Science Academy
Data Science Bootcamp Curriculum NYC Data Science Academy 100+ hours free, self-paced online course. Access to part-time in-person courses hosted at NYC campus Machine Learning with R and Python Foundations
More informationL6: Introduction to Spark Spark
L6: Introduction to Spark Spark Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Revised on December 20, 2017 Today we are going to learn...
More informationThe Definitive Guide to Preparing Your Data for Tableau
The Definitive Guide to Preparing Your Data for Tableau Speed Your Time to Visualization If you re like most data analysts today, creating rich visualizations of your data is a critical step in the analytic
More informationSpecialist ICT Learning
Specialist ICT Learning APPLIED DATA SCIENCE AND BIG DATA ANALYTICS GTBD7 Course Description This intensive training course provides theoretical and technical aspects of Data Science and Business Analytics.
More informationMachine Learning with Sparkling Water: H2O + Spark
Machine Learning with Sparkling Water: H2O + Spark Michal Malohlava Nidhi Mehta Edited by: Brandon Hill & Vinod Iyengar http://h2o.ai/resources July 2016: First Edition Machine Learning with Sparkling
More informationScalable Data Science in R and Apache Spark 2.0. Felix Cheung, Principal Engineer, Microsoft
Scalable Data Science in R and Apache Spark 2.0 Felix Cheung, Principal Engineer, Spark @ Microsoft About me Apache Spark Committer Apache Zeppelin PMC/Committer Contributing to Spark since 1.3 and Zeppelin
More informationPackage sparklyr.nested
Title A 'sparklyr' Etension for Nested Data Version 0.0.3 Package sparklyr.nested November 14, 2018 Maintainer Matt Pollock A 'sparklyr' etension adding the capability to work easily
More informationalteryx training courses
alteryx training courses alteryx designer 2 day course This course covers Alteryx Designer for new and intermediate Alteryx users. It introduces the User Interface and works through core Alteryx capability,
More informationMachine Learning with Spark. Amir H. Payberah 02/11/2018
Machine Learning with Spark Amir H. Payberah payberah@kth.se 02/11/2018 The Course Web Page https://id2223kth.github.io 1 / 89 Where Are We? 2 / 89 Where Are We? 3 / 89 Big Data 4 / 89 Problem Traditional
More informationShark: SQL and Rich Analytics at Scale. Michael Xueyuan Han Ronny Hajoon Ko
Shark: SQL and Rich Analytics at Scale Michael Xueyuan Han Ronny Hajoon Ko What Are The Problems? Data volumes are expanding dramatically Why Is It Hard? Needs to scale out Managing hundreds of machines
More informationAn Introduction to Big Data Analysis using Spark
An Introduction to Big Data Analysis using Spark Mohamad Jaber American University of Beirut - Faculty of Arts & Sciences - Department of Computer Science May 17, 2017 Mohamad Jaber (AUB) Spark May 17,
More informationBig Streaming Data Processing. How to Process Big Streaming Data 2016/10/11. Fraud detection in bank transactions. Anomalies in sensor data
Big Data Big Streaming Data Big Streaming Data Processing Fraud detection in bank transactions Anomalies in sensor data Cat videos in tweets How to Process Big Streaming Data Raw Data Streams Distributed
More informationPython, PySpark and Riak TS. Stephen Etheridge Lead Solution Architect, EMEA
Python, PySpark and Riak TS Stephen Etheridge Lead Solution Architect, EMEA Agenda Introduction to Riak TS The Riak Python client The Riak Spark connector and PySpark CONFIDENTIAL Basho Technologies 3
More information