Analyzing rtweet data with kerasformula
|
|
- Erin Davis
- 6 years ago
- Views:
Transcription
1 Analyzing rtweet data with kerasformula Pete Mohanty BARUG March 2018 hosted by Google
2 Overview Now on CRAN, kerasformula package offers a high-level interface for the R interface to Keras. install.packages("kerasformula") library(kerasformula) install_keras() # or install_keras(tensorflow = "gpu")
3 Overview Now on CRAN, kerasformula package offers a high-level interface for the R interface to Keras. install.packages("kerasformula") library(kerasformula) install_keras() # or install_keras(tensorflow = "gpu") kerasformula enables users to train and test neural nets via Keras starting with raw data in as little as one line of code.
4 Overview Now on CRAN, kerasformula package offers a high-level interface for the R interface to Keras. install.packages("kerasformula") library(kerasformula) install_keras() # or install_keras(tensorflow = "gpu") kerasformula enables users to train and test neural nets via Keras starting with raw data in as little as one line of code. This talk introduces the regression-style interface using data gathered via rtweet (developed
5 kms() kerasformula s main function Most machine learning demos assume homogenous data (e.g., pixels for digit recognition) which can make coding to fit heterogeneous data challenging. kms() takes advantage of R formulas to smooth this process.
6 kms() kerasformula s main function Most machine learning demos assume homogenous data (e.g., pixels for digit recognition) which can make coding to fit heterogeneous data challenging. kms() takes advantage of R formulas to smooth this process. kms builds dense neural nets and, after fitting them, returns a single object with predictions, measures of fit, and details about the function call.
7 kms() kerasformula s main function Most machine learning demos assume homogenous data (e.g., pixels for digit recognition) which can make coding to fit heterogeneous data challenging. kms() takes advantage of R formulas to smooth this process. kms builds dense neural nets and, after fitting them, returns a single object with predictions, measures of fit, and details about the function call. kms accepts many parameters found in keras like loss and activation functions.
8 kms() kerasformula s main function Most machine learning demos assume homogenous data (e.g., pixels for digit recognition) which can make coding to fit heterogeneous data challenging. kms() takes advantage of R formulas to smooth this process. kms builds dense neural nets and, after fitting them, returns a single object with predictions, measures of fit, and details about the function call. kms accepts many parameters found in keras like loss and activation functions. kms also accepts compiled keras_model_sequential objects allowing for even further customization.
9 The Data Let s look at #rstats tweets (excluding retweets) for a six-day period ending January 24, 2018 at 10:40. rstats <- search_tweets("#rstats", n = 10000, include_rts = FALSE) dim(rstats) [1] cor(rstats$favorite_count, rstats$retweet_count, method="spearman") [1]
10 Skewed against Popularity
11 Goal - predict popularity of each tweet rstats$popularity <- rstats$retweet_count + rstats$favorite_count Dependent variable will be discretized into approximate levels.
12 Getting the most out of formulas breaks <- c(-1, 0, 1, 10, 100, 1000, 10000) popularity <- kms(cut(popularity, breaks) ~ screen_name + source + n(hashtags) + n(mentions_screen_name) + n(urls_url) + nchar(text) + grepl('photo', media_type) + weekdays(created_at) + format(created_at, '%H'), rstats)
13 Model History
14 Confusion Matrix for First Model popularity$confusion (-1,0] (0,1] (1,10] (10,100] (100, 1000] (1000, (-1,0] (0,1] (1,10] (10,100] (100, 1000] (1000, 10000]
15 Storing the Input Formula... form <- "cut(popularity, breaks) ~ n(hashtags) + n(mentions_screen_n nchar(text) + screen_name + source + grepl('photo', media_type) + weekdays(created_at) + format(created_at, '%H')"
16 Adding Mentions Which, if any, other Twitter users are mentioned in a Tweet comes buried in variable-length list corresponding to each tweet. grepl(" ", mentions_user_id) Adding to the formula allows kms() to construct sparse model matrices directly.
17 Adding Mentions Which, if any, other Twitter users are mentioned in a Tweet comes buried in variable-length list corresponding to each tweet. grepl(" ", mentions_user_id) Adding to the formula allows kms() to construct sparse model matrices directly. M <- unique(unlist(rstats$mentions_user_id)) M <- unique(mentions[which(table(m) > 5)]) M <- M[!is.na(M)] for(m in M) form <- paste(form, "+ grepl(", m,", mentions_user_id)") popularity <- kms(pop_input, rstats)
18
19 Customizing layers with kms() Say we want to transition more gradually from the input shape to the output. l <- list(units= c(512, 256, 128, NA), activation= c("relu", "relu", "relu", "softmax"), dropout= c(0.45, 0.4, 0.35, NA)) out <- kms(pop_input, rstats, layers = l)
20 Making predictions for new data popularity <- kms(pop_input, rstats[1:1000,]) predictions <- predict(popularity, rstats[1001:2000,]) predictions$accuracy [1] 0.579
21 Making predictions for new data popularity <- kms(pop_input, rstats[1:1000,]) predictions <- predict(popularity, rstats[1001:2000,]) predictions$accuracy [1] kerasformula s predict method will automatically handle mismatches in columns between the training and the new data.
22 Using a compiled Keras model k <- keras_model_sequential() k %>% layer_embedding(input_dim = out$p, output_dim = out$p) %>% layer_lstm(units = 512, dropout = 0.4, recurrent_dropout = 0.2) %>% layer_dense(units = 256, activation = "relu") %>% layer_dropout(0.3) %>% layer_dense(units = 8, # y's number of levels activation = 'sigmoid')
23 Using a compiled Keras model k <- keras_model_sequential() k %>% layer_embedding(input_dim = out$p, output_dim = out$p) %>% layer_lstm(units = 512, dropout = 0.4, recurrent_dropout = 0.2) %>% layer_dense(units = 256, activation = "relu") %>% layer_dropout(0.3) %>% layer_dense(units = 8, # y's number of levels activation = 'sigmoid') k %>% compile( loss = 'categorical_crossentropy', optimizer = 'rmsprop', metrics = c('accuracy') )
24 Using a compiled Keras model k <- keras_model_sequential() k %>% layer_embedding(input_dim = out$p, output_dim = out$p) %>% layer_lstm(units = 512, dropout = 0.4, recurrent_dropout = 0.2) %>% layer_dense(units = 256, activation = "relu") %>% layer_dropout(0.3) %>% layer_dense(units = 8, # y's number of levels activation = 'sigmoid') k %>% compile( loss = 'categorical_crossentropy', optimizer = 'rmsprop', metrics = c('accuracy') ) out_lstm <- kms(pop_input, rstats, k)
25 Thanks! Questions? Comments? Special thanks to Dan Falbel and JJ Allaire for helpful suggestions and Joe Rickert for organizing!! Tonight s kerasformula code can be found on the RStudio Tensorflow page or the project s github repo via:
26 Choosing a Batch Size By default, kms uses batches of 32. Suppose we were happy with our model but didn t have any particular intuition about what the size should be. Nbatch <- c(16, 32, 64) Nruns <- 4 accuracy <- matrix(nrow = Nruns, ncol = length(nbatch)) colnames(accuracy) <- paste0("nbatch_", Nbatch) est <- list() for(i in 1:Nruns){ for(j in 1:length(Nbatch)){ est[[i]] <- kms(pop_input, rstats, Nepochs = 2, batch_si accuracy[i,j] <- est[[i]][["evaluations"]][["acc"]] } } colmeans(accuracy)
Package kerasformula
Package kerasformula August 23, 2018 Type Package Title A High-Level R Interface for Neural Nets Version 1.5.1 Author Pete Mohanty [aut, cre] Maintainer Pete Mohanty Description
More informationKeras: Handwritten Digit Recognition using MNIST Dataset
Keras: Handwritten Digit Recognition using MNIST Dataset IIT PATNA January 31, 2018 1 / 30 OUTLINE 1 Keras: Introduction 2 Installing Keras 3 Keras: Building, Testing, Improving A Simple Network 2 / 30
More informationKeras: Handwritten Digit Recognition using MNIST Dataset
Keras: Handwritten Digit Recognition using MNIST Dataset IIT PATNA February 9, 2017 1 / 24 OUTLINE 1 Introduction Keras: Deep Learning library for Theano and TensorFlow 2 Installing Keras Installation
More informationDeep Nets with. Keras
docs https://keras.io Deep Nets with Keras κέρας http://vem.quantumunlimited.org/the-gates-of-horn/ Professor Marie Roch These slides only cover enough to get started with feed-forward networks and do
More informationAn Introduction to Deep Learning with RapidMiner. Philipp Schlunder - RapidMiner Research
An Introduction to Deep Learning with RapidMiner Philipp Schlunder - RapidMiner Research What s in store for today?. Things to know before getting started 2. What s Deep Learning anyway? 3. How to use
More informationAn Introduction to NNs using Keras
An Introduction to NNs using Keras Michela Paganini michela.paganini@cern.ch Yale University 1 Keras Modular, powerful and intuitive Deep Learning python library built on Theano and TensorFlow Minimalist,
More informationDeep Learning. Practical introduction with Keras JORDI TORRES 27/05/2018. Chapter 3 JORDI TORRES
Deep Learning Practical introduction with Keras Chapter 3 27/05/2018 Neuron A neural network is formed by neurons connected to each other; in turn, each connection of one neural network is associated
More informationA Quick Guide on Training a neural network using Keras.
A Quick Guide on Training a neural network using Keras. TensorFlow and Keras Keras Open source High level, less flexible Easy to learn Perfect for quick implementations Starts by François Chollet from
More informationLecture 20: Neural Networks for NLP. Zubin Pahuja
Lecture 20: Neural Networks for NLP Zubin Pahuja zpahuja2@illinois.edu courses.engr.illinois.edu/cs447 CS447: Natural Language Processing 1 Today s Lecture Feed-forward neural networks as classifiers simple
More informationDeep Learning and Its Applications
Convolutional Neural Network and Its Application in Image Recognition Oct 28, 2016 Outline 1 A Motivating Example 2 The Convolutional Neural Network (CNN) Model 3 Training the CNN Model 4 Issues and Recent
More informationPackage embed. November 19, 2018
Version 0.0.2 Package embed November 19, 2018 Title Extra Recipes for Encoding Categorical Predictors Description Factor predictors can be converted to one or more numeric representations using simple
More informationDEEP LEARNING IN PYTHON. Creating a keras model
DEEP LEARNING IN PYTHON Creating a keras model Model building steps Specify Architecture Compile Fit Predict Model specification In [1]: import numpy as np In [2]: from keras.layers import Dense In [3]:
More informationPackage automl. September 13, 2018
Type Package Title Deep Learning with Metaheuristic Version 1.0.5 Author Alex Boulangé Package automl September 13, 2018 Maintainer Alex Boulangé Fits from
More informationMoonRiver: Deep Neural Network in C++
MoonRiver: Deep Neural Network in C++ Chung-Yi Weng Computer Science & Engineering University of Washington chungyi@cs.washington.edu Abstract Artificial intelligence resurges with its dramatic improvement
More informationDeep Learning Frameworks. COSC 7336: Advanced Natural Language Processing Fall 2017
Deep Learning Frameworks COSC 7336: Advanced Natural Language Processing Fall 2017 Today s lecture Deep learning software overview TensorFlow Keras Practical Graphical Processing Unit (GPU) From graphical
More informationDeep Learning with Tensorflow AlexNet
Machine Learning and Computer Vision Group Deep Learning with Tensorflow http://cvml.ist.ac.at/courses/dlwt_w17/ AlexNet Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton, "Imagenet classification
More informationTutorial on Keras CAP ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY
Tutorial on Keras CAP 6412 - ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY Deep learning packages TensorFlow Google PyTorch Facebook AI research Keras Francois Chollet (now at Google) Chainer Company
More informationDeep Learning with R. Francesca Lazzeri Data Scientist II - Microsoft, AI Research
with R Francesca Lazzeri - @frlazzeri Data Scientist II - Microsoft, AI Research Agenda with R What is Demo Better understanding of R DL tools Fundamental concepts in Forward Propagation Algorithm Activation
More informationIndex. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning,
A Acquisition function, 298, 301 Adam optimizer, 175 178 Anaconda navigator conda command, 3 Create button, 5 download and install, 1 installing packages, 8 Jupyter Notebook, 11 13 left navigation pane,
More informationFrameworks in Python for Numeric Computation / ML
Frameworks in Python for Numeric Computation / ML Why use a framework? Why not use the built-in data structures? Why not write our own matrix multiplication function? Frameworks are needed not only because
More informationMachine Learning With Python. Bin Chen Nov. 7, 2017 Research Computing Center
Machine Learning With Python Bin Chen Nov. 7, 2017 Research Computing Center Outline Introduction to Machine Learning (ML) Introduction to Neural Network (NN) Introduction to Deep Learning NN Introduction
More informationMulti-Task Self-Supervised Visual Learning
Multi-Task Self-Supervised Visual Learning Sikai Zhong March 4, 2018 COMPUTER SCIENCE Table of contents 1. Introduction 2. Self-supervised Tasks 3. Architectures 4. Experiments 1 Introduction Self-supervised
More informationLecture 8.2 Structure from Motion. Thomas Opsahl
Lecture 8.2 Structure from Motion Thomas Opsahl More-than-two-view geometry Correspondences (matching) More views enables us to reveal and remove more mismatches than we can do in the two-view case More
More informationDeep Learning Applications
October 20, 2017 Overview Supervised Learning Feedforward neural network Convolution neural network Recurrent neural network Recursive neural network (Recursive neural tensor network) Unsupervised Learning
More informationMachine Learning: Chenhao Tan University of Colorado Boulder LECTURE 15
Machine Learning: Chenhao Tan University of Colorado Boulder LECTURE 15 Slides adapted from Jordan Boyd-Graber Machine Learning: Chenhao Tan Boulder 1 of 21 Logistics HW3 available on Github, due on October
More informationKaggle Data Science Bowl 2017 Technical Report
Kaggle Data Science Bowl 2017 Technical Report qfpxfd Team May 11, 2017 1 Team Members Table 1: Team members Name E-Mail University Jia Ding dingjia@pku.edu.cn Peking University, Beijing, China Aoxue Li
More informationDEEP LEARNING IN PYTHON. Introduction to deep learning
DEEP LEARNING IN PYTHON Introduction to deep learning Imagine you work for a bank You need to predict how many transactions each customer will make next year Example as seen by linear regression Age Bank
More informationEncoding RNNs, 48 End of sentence (EOS) token, 207 Exploding gradient, 131 Exponential function, 42 Exponential Linear Unit (ELU), 44
A Activation potential, 40 Annotated corpus add padding, 162 check versions, 158 create checkpoints, 164, 166 create input, 160 create train and validation datasets, 163 dropout, 163 DRUG-AE.rel file,
More informationNeural Nets & Deep Learning
Neural Nets & Deep Learning The Inspiration Inputs Outputs Our brains are pretty amazing, what if we could do something similar with computers? Image Source: http://ib.bioninja.com.au/_media/neuron _med.jpeg
More informationDeep Neural Networks Optimization
Deep Neural Networks Optimization Creative Commons (cc) by Akritasa http://arxiv.org/pdf/1406.2572.pdf Slides from Geoffrey Hinton CSC411/2515: Machine Learning and Data Mining, Winter 2018 Michael Guerzhoy
More informationNatural Language Processing with Deep Learning CS224N/Ling284. Christopher Manning Lecture 4: Backpropagation and computation graphs
Natural Language Processing with Deep Learning CS4N/Ling84 Christopher Manning Lecture 4: Backpropagation and computation graphs Lecture Plan Lecture 4: Backpropagation and computation graphs 1. Matrix
More informationSEMANTIC COMPUTING. Lecture 8: Introduction to Deep Learning. TU Dresden, 7 December Dagmar Gromann International Center For Computational Logic
SEMANTIC COMPUTING Lecture 8: Introduction to Deep Learning Dagmar Gromann International Center For Computational Logic TU Dresden, 7 December 2018 Overview Introduction Deep Learning General Neural Networks
More information[3-5] Consider a Combiner that tracks local counts of followers and emits only the local top10 users with their number of followers.
Quiz 6 We design of a MapReduce application that generates the top-10 most popular Twitter users (based on the number of followers) of a month based on the number of followers. Suppose that you have the
More informationPractical 8: Neural networks
Practical 8: Neural networks Properly building and training a neural network involves many design decisions such as choosing number and nature of layers and fine-tuning hyperparameters. Many details are
More informationGPU Based Image Classification using Convolutional Neural Network Chicken Dishes Classification
Int. J. Advance Soft Compu. Appl, Vol. 10, No. 2, July 2018 ISSN 2074-8523 GPU Based Image Classification using Convolutional Neural Network Chicken Dishes Classification Hafizhan Aliady, Dina Tri Utari
More informationNatural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu
Natural Language Processing CS 6320 Lecture 6 Neural Language Models Instructor: Sanda Harabagiu In this lecture We shall cover: Deep Neural Models for Natural Language Processing Introduce Feed Forward
More informationNeural networks. About. Linear function approximation. Spyros Samothrakis Research Fellow, IADS University of Essex.
Neural networks Spyros Samothrakis Research Fellow, IADS University of Essex About Linear function approximation with SGD From linear regression to neural networks Practical aspects February 28, 2017 Conclusion
More informationDeep Learning on Graphs
Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 22 March 2017 joint work with Max Welling (University of Amsterdam) BDL Workshop @ NIPS 2016
More informationPOINT CLOUD DEEP LEARNING
POINT CLOUD DEEP LEARNING Innfarn Yoo, 3/29/28 / 57 Introduction AGENDA Previous Work Method Result Conclusion 2 / 57 INTRODUCTION 3 / 57 2D OBJECT CLASSIFICATION Deep Learning for 2D Object Classification
More informationAutomation.
Automation www.austech.edu.au WHAT IS AUTOMATION? Automation testing is a technique uses an application to implement entire life cycle of the software in less time and provides efficiency and effectiveness
More informationDeep Learning Basic Lecture - Complex Systems & Artificial Intelligence 2017/18 (VO) Asan Agibetov, PhD.
Deep Learning 861.061 Basic Lecture - Complex Systems & Artificial Intelligence 2017/18 (VO) Asan Agibetov, PhD asan.agibetov@meduniwien.ac.at Medical University of Vienna Center for Medical Statistics,
More informationarxiv: v1 [stat.ml] 21 Feb 2018
Detecting Learning vs Memorization in Deep Neural Networks using Shared Structure Validation Sets arxiv:2.0774v [stat.ml] 2 Feb 8 Elias Chaibub Neto e-mail: elias.chaibub.neto@sagebase.org, Sage Bionetworks
More informationRecurrent neural networks for polyphonic sound event detection in real life recordings
Recurrent neural networks for polyphonic sound event detection in real life recordings Giambattista Parascandolo, Heikki Huttunen, Tuomas Virtanen Tampere University of Technology giambattista.parascandolo@tut.fi
More informationAdvanced Machine Learning
Advanced Machine Learning Convolutional Neural Networks for Handwritten Digit Recognition Andreas Georgopoulos CID: 01281486 Abstract Abstract At this project three different Convolutional Neural Netwroks
More informationConvolutional Neural Networks
Lecturer: Barnabas Poczos Introduction to Machine Learning (Lecture Notes) Convolutional Neural Networks Disclaimer: These notes have not been subjected to the usual scrutiny reserved for formal publications.
More informationCPSC 340: Machine Learning and Data Mining. Deep Learning Fall 2016
CPSC 340: Machine Learning and Data Mining Deep Learning Fall 2016 Assignment 5: Due Friday. Assignment 6: Due next Friday. Final: Admin December 12 (8:30am HEBB 100) Covers Assignments 1-6. Final from
More informationDeep Learning for Computer Vision
Deep Learning for Computer Vision Lecture 7: Universal Approximation Theorem, More Hidden Units, Multi-Class Classifiers, Softmax, and Regularization Peter Belhumeur Computer Science Columbia University
More informationCOMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017
COMP9444 Neural Networks and Deep Learning 7. Image Processing COMP9444 17s2 Image Processing 1 Outline Image Datasets and Tasks Convolution in Detail AlexNet Weight Initialization Batch Normalization
More informationMachine Learning Workshop
Machine Learning Workshop {Presenters} Feb. 20th, 2018 Theory of Neural Networks Architecture and Types of Layers: Fully Connected (FC) Convolutional Neural Network (CNN) Pooling Drop out Residual Recurrent
More informationMachine Learning 13. week
Machine Learning 13. week Deep Learning Convolutional Neural Network Recurrent Neural Network 1 Why Deep Learning is so Popular? 1. Increase in the amount of data Thanks to the Internet, huge amount of
More informationDeep Learning for Computer Vision with MATLAB By Jon Cherrie
Deep Learning for Computer Vision with MATLAB By Jon Cherrie 2015 The MathWorks, Inc. 1 Deep learning is getting a lot of attention "Dahl and his colleagues won $22,000 with a deeplearning system. 'We
More informationOnion Language Reference Manual. By: Andrew Aday, (aza2112) Amol Kapoor (ajk2227), Jonathan Zhang (jz2814)
Onion Language Reference Manual By: Andrew Aday, (aza2112) Amol Kapoor (ajk2227), Jonathan Zhang (jz2814) Introduction 3 Types 3 Lexical Conventions 4 Identifiers 4 Keywords 4 Comments 4 Expressions 5
More informationAccelerating Convolutional Neural Nets. Yunming Zhang
Accelerating Convolutional Neural Nets Yunming Zhang Focus Convolutional Neural Nets is the state of the art in classifying the images The models take days to train Difficult for the programmers to tune
More informationPackage tensorflow. January 17, 2018
Type Package Title R Interface to 'TensorFlow' Version 1.5 Package tensorflow January 17, 2018 Interface to 'TensorFlow' , an open source software library for numerical computation
More informationFace Recognition A Deep Learning Approach
Face Recognition A Deep Learning Approach Lihi Shiloh Tal Perl Deep Learning Seminar 2 Outline What about Cat recognition? Classical face recognition Modern face recognition DeepFace FaceNet Comparison
More informationConverting categorical data into numbers with Pandas and Scikit-learn -...
1 of 6 11/17/2016 11:02 AM FastML Machine learning made easy RSS Home Contents Popular Links Backgrounds About Converting categorical data into numbers with Pandas and Scikit-learn 2014-04-30 Many machine
More informationProgramming Exercise 4: Neural Networks Learning
Programming Exercise 4: Neural Networks Learning Machine Learning Introduction In this exercise, you will implement the backpropagation algorithm for neural networks and apply it to the task of hand-written
More informationDeep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group
Deep Learning Vladimir Golkov Technical University of Munich Computer Vision Group 1D Input, 1D Output target input 2 2D Input, 1D Output: Data Distribution Complexity Imagine many dimensions (data occupies
More informationInference Optimization Using TensorRT with Use Cases. Jack Han / 한재근 Solutions Architect NVIDIA
Inference Optimization Using TensorRT with Use Cases Jack Han / 한재근 Solutions Architect NVIDIA Search Image NLP Maps TensorRT 4 Adoption Use Cases Speech Video AI Inference is exploding 1 Billion Videos
More informationTechnical University of Munich. Exercise 8: Neural Networks
Technical University of Munich Chair for Bioinformatics and Computational Biology Protein Prediction I for Computer Scientists SoSe 2018 Prof. B. Rost M. Bernhofer, M. Heinzinger, D. Nechaev, L. Richter
More informationObject Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal
Object Detection Lecture 10.3 - Introduction to deep learning (CNN) Idar Dyrdal Deep Learning Labels Computational models composed of multiple processing layers (non-linear transformations) Used to learn
More informationReal-time Scheduling of Skewed MapReduce Jobs in Heterogeneous Environments
Real-time Scheduling of Skewed MapReduce Jobs in Heterogeneous Environments Nikos Zacheilas, Vana Kalogeraki Department of Informatics Athens University of Economics and Business 1 Big Data era has arrived!
More informationUsing Neural Cells to Improve Image Textual Line Segmentation
Using Neural Cells to Improve Image Textual Line Segmentation Patrick Schone (patrickjohn.schone@ldschurch.org) 7 February 2017 Standards Technical Conference Overview Motivation Neural Cells for Line
More informationMachine Learning and Algorithms for Data Mining Practical 2: Neural Networks
CST Part III/MPhil in Advanced Computer Science 2016-2017 Machine Learning and Algorithms for Data Mining Practical 2: Neural Networks Demonstrators: Petar Veličković, Duo Wang Lecturers: Mateja Jamnik,
More informationPractical Deep Learning
Practical Deep Learning micha.codes / fastforwardlabs.com 1 / 70 deep learning can seem mysterious 2 / 70 let's nd a way to just build a function 3 / 70 Feed Forward Layer # X.shape == (512,) # output.shape
More informationECE 5470 Classification, Machine Learning, and Neural Network Review
ECE 5470 Classification, Machine Learning, and Neural Network Review Due December 1. Solution set Instructions: These questions are to be answered on this document which should be submitted to blackboard
More informationIntroduction to R: Part I
Introduction to R: Part I Jeffrey C. Miecznikowski March 26, 2015 R impact R is the 13th most popular language by IEEE Spectrum (2014) Google uses R for ROI calculations Ford uses R to improve vehicle
More informationApparel Classification using CNNs
Apparel Classification using CNNs Rohit Patki ICME Stanford University rpatki@stanford.edu Suhas Suresha ICME Stanford University suhas17@stanford.edu Abstract Apparel classification from images finds
More informationDeveloping Machine Learning Models. Kevin Tsai
Developing Machine Learning Models Kevin Tsai GOTO Chicago 2018 Developing Machine Learning Models Kevin Tsai, Google Agenda Machine Learning with tf.estimator Performance pipelines with TFRecords and
More informationApplication of Deep Learning Techniques in Satellite Telemetry Analysis.
Application of Deep Learning Techniques in Satellite Telemetry Analysis. Greg Adamski, Member of Technical Staff L3 Technologies Telemetry and RF Products Julian Spencer Jones, Spacecraft Engineer Telenor
More informationINTRODUCTION TO DEEP LEARNING
INTRODUCTION TO DEEP LEARNING CONTENTS Introduction to deep learning Contents 1. Examples 2. Machine learning 3. Neural networks 4. Deep learning 5. Convolutional neural networks 6. Conclusion 7. Additional
More informationReal-time Gesture Pattern Classification with IMU Data
Real-time Gesture Pattern Classification with IMU Data Alex Fu Stanford University Computer Science Department alexfu@stanford.edu Yangyang Yu Stanford University Electrical Engineering Department yyu10@stanford.edu
More informationExploring the Structure of Data at Scale. Rudy Agovic, PhD CEO & Chief Data Scientist at Reliancy January 16, 2019
Exploring the Structure of Data at Scale Rudy Agovic, PhD CEO & Chief Data Scientist at Reliancy January 16, 2019 Outline Why exploration of large datasets matters Challenges in working with large data
More informationNeural Networks with Input Specified Thresholds
Neural Networks with Input Specified Thresholds Fei Liu Stanford University liufei@stanford.edu Junyang Qian Stanford University junyangq@stanford.edu Abstract In this project report, we propose a method
More informationConvolutional Layer Pooling Layer Fully Connected Layer Regularization
Semi-Parallel Deep Neural Networks (SPDNN), Convergence and Generalization Shabab Bazrafkan, Peter Corcoran Center for Cognitive, Connected & Computational Imaging, College of Engineering & Informatics,
More informationDeep Learning. Volker Tresp Summer 2017
Deep Learning Volker Tresp Summer 2017 1 Neural Network Winter and Revival While Machine Learning was flourishing, there was a Neural Network winter (late 1990 s until late 2000 s) Around 2010 there
More informationDeep neural networks II
Deep neural networks II May 31 st, 2018 Yong Jae Lee UC Davis Many slides from Rob Fergus, Svetlana Lazebnik, Jia-Bin Huang, Derek Hoiem, Adriana Kovashka, Why (convolutional) neural networks? State of
More informationMulti-layer Perceptron Forward Pass Backpropagation. Lecture 11: Aykut Erdem November 2016 Hacettepe University
Multi-layer Perceptron Forward Pass Backpropagation Lecture 11: Aykut Erdem November 2016 Hacettepe University Administrative Assignment 2 due Nov. 10, 2016! Midterm exam on Monday, Nov. 14, 2016 You are
More informationConvolutional Neural Network for Facial Expression Recognition
Convolutional Neural Network for Facial Expression Recognition Liyuan Zheng Department of Electrical Engineering University of Washington liyuanz8@uw.edu Shifeng Zhu Department of Electrical Engineering
More informationPointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space
PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space Sikai Zhong February 14, 2018 COMPUTER SCIENCE Table of contents 1. PointNet 2. PointNet++ 3. Experiments 1 PointNet Property
More informationBusiness Club. Decision Trees
Business Club Decision Trees Business Club Analytics Team December 2017 Index 1. Motivation- A Case Study 2. The Trees a. What is a decision tree b. Representation 3. Regression v/s Classification 4. Building
More informationConvolutional Neural Networks for Facial Expression Recognition
Convolutional Neural Networks for Facial Expression Recognition Shima Alizadeh Stanford University shima86@stanford.edu Azar Fazel Stanford University azarf@stanford.edu Abstract In this project, we have
More informationConvolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech
Convolutional Neural Networks Computer Vision Jia-Bin Huang, Virginia Tech Today s class Overview Convolutional Neural Network (CNN) Training CNN Understanding and Visualizing CNN Image Categorization:
More informationSoftware Defined Hardware
Software Defined Hardware For data intensive computation Wade Shen DARPA I2O September 19, 2017 1 Goal Statement Build runtime reconfigurable hardware and software that enables near ASIC performance (within
More information3D model classification using convolutional neural network
3D model classification using convolutional neural network JunYoung Gwak Stanford jgwak@cs.stanford.edu Abstract Our goal is to classify 3D models directly using convolutional neural network. Most of existing
More informationImage Classification using Deep Learning with Support Vector Machines
Machine Learning Engineer Nanodegree Capstone Project Report Pavlos Sakoglou November 23, 2017 Image Classification using Deep Learning with Support Vector Machines 1 Definition 1.1 Domain Background Convolutional
More informationCS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes
CS 6501: Deep Learning for Computer Graphics Training Neural Networks II Connelly Barnes Overview Preprocessing Initialization Vanishing/exploding gradients problem Batch normalization Dropout Additional
More informationDeep Convolutional Neural Networks. Nov. 20th, 2015 Bruce Draper
Deep Convolutional Neural Networks Nov. 20th, 2015 Bruce Draper Background: Fully-connected single layer neural networks Feed-forward classification Trained through back-propagation Example Computer Vision
More informationCLASSIFICATION JELENA JOVANOVIĆ. Web:
CLASSIFICATION JELENA JOVANOVIĆ Email: jeljov@gmail.com Web: http://jelenajovanovic.net OUTLINE What is classification? Binary and multiclass classification Classification algorithms Naïve Bayes (NB) algorithm
More informationAll You Want To Know About CNNs. Yukun Zhu
All You Want To Know About CNNs Yukun Zhu Deep Learning Deep Learning Image from http://imgur.com/ Deep Learning Image from http://imgur.com/ Deep Learning Image from http://imgur.com/ Deep Learning Image
More informationCode Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python:
Code Mania 2019 Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python: 1. Introduction to Artificial Intelligence 2. Introduction to python programming and Environment
More informationR for SQListas, a Continuation
3-2 - 1-0: Classifying Digits with R R for SQListas, a Continuation R for SQListas: Now that we're in the tidyverse... what can we do now? Machine Learning MNIST - the Drosophila of Machine Learning (attributed
More informationClassifying Depositional Environments in Satellite Images
Classifying Depositional Environments in Satellite Images Alex Miltenberger and Rayan Kanfar Department of Geophysics School of Earth, Energy, and Environmental Sciences Stanford University 1 Introduction
More informationCS1114 Assignment 5, Part 1
CS4 Assignment 5, Part out: Friday, March 27, 2009. due: Friday, April 3, 2009, 5PM. This assignment covers three topics in two parts: interpolation and image transformations (Part ), and feature-based
More informationEECS 496 Statistical Language Models. Winter 2018
EECS 496 Statistical Language Models Winter 2018 Introductions Professor: Doug Downey Course web site: www.cs.northwestern.edu/~ddowney/courses/496_winter2018 (linked off prof. home page) Logistics Grading
More informationReview: The best frameworks for machine learning and deep learning
Review: The best frameworks for machine learning and deep learning infoworld.com/article/3163525/analytics/review-the-best-frameworks-for-machine-learning-and-deep-learning.html By Martin Heller Over the
More informationCounting Passenger Vehicles from Satellite Imagery
Counting Passenger Vehicles from Satellite Imagery Not everything that can be counted counts, and not everything that counts can be counted NVIDIA GPU Technology Conference 02 Nov 2017 KEVIN GREEN MACHINE
More informationConvolution Neural Networks for Chinese Handwriting Recognition
Convolution Neural Networks for Chinese Handwriting Recognition Xu Chen Stanford University 450 Serra Mall, Stanford, CA 94305 xchen91@stanford.edu Abstract Convolutional neural networks have been proven
More informationEE 511 Neural Networks
Slides adapted from Ali Farhadi, Mari Ostendorf, Pedro Domingos, Carlos Guestrin, and Luke Zettelmoyer, Andrei Karpathy EE 511 Neural Networks Instructor: Hanna Hajishirzi hannaneh@washington.edu Computational
More informationRNN LSTM and Deep Learning Libraries
RNN LSTM and Deep Learning Libraries UDRC Summer School Muhammad Awais m.a.rana@surrey.ac.uk Outline Recurrent Neural Network Application of RNN LSTM Caffe Torch Theano TensorFlow Flexibility of Recurrent
More information