Facial Expression Classification with Random Filters Feature Extraction
|
|
- Darcy Payne
- 6 years ago
- Views:
Transcription
1 Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle the challenging problem of facial expression classification. We were assigned with images of faces labeled with seven categories and unlabeled faces from the Toronto Faces Dataset [1]. We designed a machine learning model using random filters in the convolutional layer, followed by a ReLu and a max pooling layer, and a linear SVM to discriminate between classes. We achieved a very high classification rate of 84.7% in the public test set (418 images) and 83.7% in the private test set (835 images), ranked first and second respectively among 60 teams. II. BACKGROUND Convolutional neural networks are very successful models in the image object classification tasks [2]. Those networks usually have a convolutional layer with shared weights in the first layer, which can be viewed as image filters. Filtered images are then passed to a Rectified Linear Unit (ReLU) to obtain non linear activation to certain features. Next, to reduce data dimensionality and to make the model robust to translational invariance, a max pooling layer effectively downsamples the intermediate image with the maximum value in each subregion. Lastly, a fully connected softmax layer makes class predictions with each unit representing the probability of input data belonging to certain classes. Recent works show that for small number of total classes, a set of one vs all support vector machines (SVM) can replace the fully connected layers to reduce the model training complexity and achieve comparable results [7]. Moreover, a better model architecture (with ReLU and pooling) seems to be more important than the pretraining of the weights in the convolutional filters. In fact, research shows that randomly generated filters can perform surprisingly well [6]. The reason is that those filters provide an overcomplete set of bases that map the original image to a high dimensional space [8]. The ReLU and pooling layer increase data sparsity and preserve the most salient features in the image, and these features can then be separated by a linear SVM. III. MODEL Our model consists of a convolutional layer of random weights, a ReLU (positive and negative direction), and a max pooling layer. We then send the processed data into a linear SVM. We compared our model with other set of common filters such as gabor filters and sparse coding filters, and other popular methods such as knn, k means SVM, random forest and logistics. We discovered that the random filters perform the best among all models that we have tried. To achieve better results, we also experimented with different hyperparameters such as filter size, number of filters, pooling region, hierarchical pooling. We show that a set of 1024 randomly generated 8x8 filters with hierarchical pooling performs the best among single classifiers. Page 1 of 7
2 Moreover, we explored bagging method and constructed an ensemble of 15 classifiers with 40 filters, which increased the classification rate by 1%. Preprocessing First we preprocess the data to ensure each patch has zero mean and uni variance. Alternatively, we whiten the patches. ZCA Whitening transforms the data to have zero mean and an identity covariance matrix. Convolutional Layer Given F P P filters, we perform convolution on the original M N image. The valid region of the convolution output is an image of size ( M P + 1 ) ( N P + 1 ) = M N. For each filter output s, we pass it to a ReLU function: z + = m ax(0, s). In order to respond to negative features, we also have a negative ReLU: z = max(0, s ). In previous work of image object classification, positive and negative ReLU pairs are shown to be effective[3]. The ReLU suppresses the value to zero for non positive values, thus creates data sparsity for better data separation. We concatenate the features from both positive and negative ReLU functions, obtaining 2 F filtered images. After the ReLU, the model gives 2 F M N images. We then downsample the image into 2 F m n smaller images by selecting the largest value in the M /m N /n region. Alternatively, we can also use different downsample function since max pooling is in fact the L norm of the pooling region. L 1 norm corresponds to sum pooling. We are free to choose any parameter p to compute the p norm of the region as the pooled value. Lastly, we flatten the 2 F m n matrix into an image feature vector, to pass into an SVM with linear kernel. In order to classify between seven classes, we trained seven 1 vs all SVMs. The 1 following diagrams show the entire architecture of our proposed model. Figure 1 Our Proposed Model Architecture 1 Passing the filtered data into a positive and negative ReLU respectively results in 2F sets of filtered data. Page 2 of 7
3 We plot the images of the filters and the activated, downsampled image of a face. Figure 2 Random Filters (Left). 5x5 Activation after ReLU and Max Pooling (Right). As we can see, there is no particular pattern in the filter initialization; however, they seem to be sensitive to regions such as eyes, noses and mouth, which explains why random filters are effective. The grid like structure in the filter images resembles a small high pass image filter, which detects high frequency variation region such as facial features and ignores low variation region such as different skin colours. Classification Layer In the classification layer, we have seven one vs. all SVM classifiers. Each gives prediction of whether an image belongs to a class, as well as the distance to the decision boundary. We take the longest distance among all classifiers on the positive side of the decision boundary as the final prediction of an image. IV. EXPERIMENT Throughout the project, we experimented with a variety of approaches and methods. In particular, we experimented with different filters, pooling regions, and ensemble methods. The final model is chosen according to the analysis of all these experiments. In this process, besides testing hyper parameters for random filters, we examined the effectiveness of bagging; we also attempted to learn a gating function using a mixture of experts. Moreover, preprocessing using k means clusters, are also tried. In the end, they were not selected in the final model as their performance does not match the selected model and they cannot be used to help the existing model. The detailed experiments results are discussed below. Filter Size We found that 8 8 filters work the best for both random and Gabor filters. They actually capture an entire eye and nose. Page 3 of 7
4 Effect of Different Filters We experimented different filters to extract new image representation. Besides randomly generated filters, we also experimented with Gabor filters[9], sparse coding[3], and k means coding[4]. Gabor filters are famous for edge detection [4]. Sparse coding and k means coding are popular encoding techniques to map data into sparse high dimensional space. Figure 3 Gabor Filters (Left) and Sparse Coding Dictionary (Right) 1024 Random Filters (no whiten) 40 Random Filters (no whiten) 4x10 Gabor Filters (whiten) 40 Sparse Coding Dictionary (whiten) [2] 40 K means Cluster Triangular Encoding 2 (no whiten) [3] 2 Fold Rate 79.61% 76.34% 77.01% 73.33% 49.47% Table 1 Effect of Different Filters Gabor filters and sparse coding filters perform better in whitened data, whereas random filters perform better by simply shifting mean and divide by standard deviation. Gabor filters have a slight advantage over random filters given the same number of filters, but there is less space for improvement because the shape and form of the filters are rather restricted. With 1024 random filters, we achieved a classification rate of 80% with only half of the data (1460 labeled images). Number of Filters As shown above, 40 random filters can achieve reasonably good classification rate. The benefit of more filters is exponentially decaying filters increased the classification rate by 3%. 3 Pooling Regions Random Filters 2x2 Square 5x5 Square Squares Squares + Rectangles 6 Held out rate 75.0% 81.8% 83.01% 84.21% Table 2 Effect of Different Pooling Regions 2 triangular K means follows the method outlined in [4], a version of soft k means that is able to keep some sparsity. 3 m x n pooling means dissecting images into non overlapping m x n regions, and sample one value in each region. 4 We used 5x5, 4x4, 3x3, 2x2 squares and concatenated the features into feature vector. 5 On top of the square from 5x5 to 2x2, we used rectangular pooling region of 5x1, 4x1, 3x1, and 2x1. 6 The held out set composed of 418 images, available as public test set on Kaggle. Page 4 of 7
5 To select a better architecture, we experimented with different pooling regions. The above table summarizes how different pooling mechanism affect the classification rate. 2x2 square max pooling suffers from loss of information. We also discover that it is beneficial to perform pooling on different sizes of pooling region. With both hierarchical square pooling boxes and rectangular pooling boxes, we achieved a classification rate of 84.21% on the held out set. Max pooling or L p norm pooling Although max pooling may be prone to loss of local features, we found that L p norm (p = 5, 10, 20) pooling does not outperform max pooling on the held out set. The loss of information can be mitigated by more hierarchical pooling regions. Training Time One of the biggest advantages of random filters is that it does not need any pre training for obtaining the filters. We list a comparison of training time. 512 Random Filters (RF) 512 Gabor Filters (GF) 512 Sparse 7 Coding (SC) K means Unsupervised Training Phase hours 0.5 hour Table 3 Unsupervised Training Time Comparison Between Encoding Techniques. In the supervised phase, SVM can train 3000 images in 5 minutes. The tradeoff of SVM is the growth of the model size. For 1024 random filters with hierarchical square and rectangle pooling regions, the trained model is approximately 3 GB. SVM Kernel Selection We found that linear kernel actually performs the best. This suggests that in our high dimensional image representation space, categories are linearly separable. Polynomial kernel does not produce any meaningful results; RBF kernel with sigma equal to 25 produces a 2 fold classification rate of 66%, which is far below linear kernel. Ensemble Method We also experimented with an ensemble of classifiers with random filters, gabor filters, and sparse coding filters. The final prediction is the arithmetic average of the prediction probability of all classifiers in the ensemble. Bag Ratio 0.8 Single 512 RF 15x40 RF 20x40 GF 10xRF+5xGF+5xSC Held out rate 83.01% 84.68% 82.54% 82.30% Table 4 Ensemble Method Results 7 We used stochastic coordinate gradient descent with mini batch of We used fkmeans (fast kmeans) from MATLABCentral by Tim Benham to boost speed [5] Page 5 of 7
6 We discover that with an ensemble of 15 classifiers (600 filters in total), we get 1.7% increase in classification rate compared to a single classifier with similar number of filters. Comparison with Other Methods a. K Nearest Neighbours (KNN), Bag of KNN KNN algorithm is provided in the project starter package. It is a baseline classifier in this task. On top of the baseline KNN, we also built an ensemble of KNN using bagging. b. Random Forest We first obtain the major 45 principal components (PCA), and then train a bag of 7x200 1 vs. all decision trees using MATLAB function fitensemble. c. Polynomial SVM We first obtain the major 45 PCA components, and then train a 7x one vs. all SVM with 3rd order polynomial kernel. d. Multinomial Logistic Regression (LR) We first pass the images through random filters, then we train 7x one vs. all SVM with linear kernel RF KNN (K=5) Bag KNN Random Forest Poly SVM Multi LR 10 Fold Rate 82.56% 57.39% 61.03% 64.20% 65.13% 67.57% Held out Rate 84.21% 58.12% Table 5 10 Fold Cross Validation Comparison of Different Classifiers. Most of the methods we explored beat the baseline KNN. The 1024 Random Filters beats the baseline by more than 25%. Held out rates of other methods are not provided because of the submission limit. V. CONCLUSION According to the abovementioned experiments and the laid out analysis, we found our proposed model works best. We obtain new representation of data by convolving random filters with original images. The positive and negative features are extracted respectively by passing the filtered data through a positive ReLU and a negative ReLU. These data from all filters are then pooled to create a new vector that becomes the new representation of the original data. To classify the new representations, 7 one vs. all linear SVMs are trained and used. Finally, it is proven effective to train multiple SVMs using different sets of random filters; the average result of these separate SVM classifiers can further boost the overall classification rate. Page 6 of 7
7 REFERENCES [1] The Toronto Faces Dataset, [2] LeCun et al. Gradient based learning applied to document recognition. In Proceedings of the IEEE, [3] Adam Coates and Andrew Y. Ng, The Importance of Encoding Versus Training with Sparse Coding and Vector Quantization. In ICML , [4] Adam Coates, Honglak Lee and Andrew Y. Ng, An Analysis of Single Layer Networks in Unsupervised Feature Learning. In 14th ICAIS [5] Tim Benham, Fast K means Implementation With Optional Weights, fast k means [6] Andrew M. Saxe et al. On Random Weights and Unsupervised Feature Learning. In ICML 11, [7] Yichuan Tang, Deep Learning using Linear Support Vector Machines. ICML 13, [8] B. Olshausen and D. Field, Sparse Coding with an Overcomplete Basis Set: A Strategy Employed by V1? Vision Research, vol. 37, [9] M. Haghighat, S. Zonouz, M. Abdel Mottaleb, "Identification Using Encrypted Biometrics," Computer Analysis of Images and Patterns, Springer Berlin Heidelberg, pp , Page 7 of 7
Tutorial on Machine Learning Tools
Tutorial on Machine Learning Tools Yanbing Xue Milos Hauskrecht Why do we need these tools? Widely deployed classical models No need to code from scratch Easy-to-use GUI Outline Matlab Apps Weka 3 UI TensorFlow
More informationCOMP 551 Applied Machine Learning Lecture 16: Deep Learning
COMP 551 Applied Machine Learning Lecture 16: Deep Learning Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted, all
More informationHuman Vision Based Object Recognition Sye-Min Christina Chan
Human Vision Based Object Recognition Sye-Min Christina Chan Abstract Serre, Wolf, and Poggio introduced an object recognition algorithm that simulates image processing in visual cortex and claimed to
More informationCOSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor
COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality
More informationConvolution Neural Networks for Chinese Handwriting Recognition
Convolution Neural Networks for Chinese Handwriting Recognition Xu Chen Stanford University 450 Serra Mall, Stanford, CA 94305 xchen91@stanford.edu Abstract Convolutional neural networks have been proven
More informationMulti-Task Learning of Facial Landmarks and Expression
Multi-Task Learning of Facial Landmarks and Expression Terrance Devries 1, Kumar Biswaranjan 2, and Graham W. Taylor 1 1 School of Engineering, University of Guelph, Guelph, Canada N1G 2W1 2 Department
More informationSparse coding for image classification
Sparse coding for image classification Columbia University Electrical Engineering: Kun Rong(kr2496@columbia.edu) Yongzhou Xiang(yx2211@columbia.edu) Yin Cui(yc2776@columbia.edu) Outline Background Introduction
More informationNeural Network Optimization and Tuning / Spring 2018 / Recitation 3
Neural Network Optimization and Tuning 11-785 / Spring 2018 / Recitation 3 1 Logistics You will work through a Jupyter notebook that contains sample and starter code with explanations and comments throughout.
More informationIndex. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning,
A Acquisition function, 298, 301 Adam optimizer, 175 178 Anaconda navigator conda command, 3 Create button, 5 download and install, 1 installing packages, 8 Jupyter Notebook, 11 13 left navigation pane,
More informationMachine Learning 13. week
Machine Learning 13. week Deep Learning Convolutional Neural Network Recurrent Neural Network 1 Why Deep Learning is so Popular? 1. Increase in the amount of data Thanks to the Internet, huge amount of
More informationPreface to the Second Edition. Preface to the First Edition. 1 Introduction 1
Preface to the Second Edition Preface to the First Edition vii xi 1 Introduction 1 2 Overview of Supervised Learning 9 2.1 Introduction... 9 2.2 Variable Types and Terminology... 9 2.3 Two Simple Approaches
More informationOn Classification: An Empirical Study of Existing Algorithms Based on Two Kaggle Competitions
On Classification: An Empirical Study of Existing Algorithms Based on Two Kaggle Competitions CAMCOS Report Day December 9th, 2015 San Jose State University Project Theme: Classification The Kaggle Competition
More informationLearning to Recognize Faces in Realistic Conditions
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
More informationAdvanced Machine Learning
Advanced Machine Learning Convolutional Neural Networks for Handwritten Digit Recognition Andreas Georgopoulos CID: 01281486 Abstract Abstract At this project three different Convolutional Neural Netwroks
More informationDeep Learning for Computer Vision II
IIIT Hyderabad Deep Learning for Computer Vision II C. V. Jawahar Paradigm Shift Feature Extraction (SIFT, HoG, ) Part Models / Encoding Classifier Sparrow Feature Learning Classifier Sparrow L 1 L 2 L
More informationExploring Bag of Words Architectures in the Facial Expression Domain
Exploring Bag of Words Architectures in the Facial Expression Domain Karan Sikka, Tingfan Wu, Josh Susskind, and Marian Bartlett Machine Perception Laboratory, University of California San Diego {ksikka,ting,josh,marni}@mplab.ucsd.edu
More informationDivya Ramesh December 8, 2014
EE 660 Machine Learning from Signals: Foundations and Methods Project Title: Analysis of Unsupervised Feature Learning Methods for Image Classification Divya Ramesh (dramesh@usc.edu) 5924-4440-58 December
More informationDeep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.
Visualizing and Understanding Convolutional Networks Christopher Pennsylvania State University February 23, 2015 Some Slide Information taken from Pierre Sermanet (Google) presentation on and Computer
More informationContents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation
Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation Learning 4 Supervised Learning 4 Unsupervised Learning 4
More informationBayes Risk. Classifiers for Recognition Reading: Chapter 22 (skip 22.3) Discriminative vs Generative Models. Loss functions in classifiers
Classifiers for Recognition Reading: Chapter 22 (skip 22.3) Examine each window of an image Classify object class within each window based on a training set images Example: A Classification Problem Categorize
More informationLecture 27: Review. Reading: All chapters in ISLR. STATS 202: Data mining and analysis. December 6, 2017
Lecture 27: Review Reading: All chapters in ISLR. STATS 202: Data mining and analysis December 6, 2017 1 / 16 Final exam: Announcements Tuesday, December 12, 8:30-11:30 am, in the following rooms: Last
More informationClassifiers for Recognition Reading: Chapter 22 (skip 22.3)
Classifiers for Recognition Reading: Chapter 22 (skip 22.3) Examine each window of an image Classify object class within each window based on a training set images Slide credits for this chapter: Frank
More informationConvolutional-Recursive Deep Learning for 3D Object Classification
Convolutional-Recursive Deep Learning for 3D Object Classification Richard Socher, Brody Huval, Bharath Bhat, Christopher D. Manning, Andrew Y. Ng NIPS 2012 Iro Armeni, Manik Dhar Motivation Hand-designed
More informationPredicting Popular Xbox games based on Search Queries of Users
1 Predicting Popular Xbox games based on Search Queries of Users Chinmoy Mandayam and Saahil Shenoy I. INTRODUCTION This project is based on a completed Kaggle competition. Our goal is to predict which
More informationComparison of different preprocessing techniques and feature selection algorithms in cancer datasets
Comparison of different preprocessing techniques and feature selection algorithms in cancer datasets Konstantinos Sechidis School of Computer Science University of Manchester sechidik@cs.man.ac.uk Abstract
More informationCPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016
CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2016 A2/Midterm: Admin Grades/solutions will be posted after class. Assignment 4: Posted, due November 14. Extra office hours:
More informationThe exam is closed book, closed notes except your one-page (two-sided) cheat sheet.
CS 189 Spring 2015 Introduction to Machine Learning Final You have 2 hours 50 minutes for the exam. The exam is closed book, closed notes except your one-page (two-sided) cheat sheet. No calculators or
More informationClassifying Depositional Environments in Satellite Images
Classifying Depositional Environments in Satellite Images Alex Miltenberger and Rayan Kanfar Department of Geophysics School of Earth, Energy, and Environmental Sciences Stanford University 1 Introduction
More informationLinear Discriminant Analysis for 3D Face Recognition System
Linear Discriminant Analysis for 3D Face Recognition System 3.1 Introduction Face recognition and verification have been at the top of the research agenda of the computer vision community in recent times.
More informationCS 179 Lecture 16. Logistic Regression & Parallel SGD
CS 179 Lecture 16 Logistic Regression & Parallel SGD 1 Outline logistic regression (stochastic) gradient descent parallelizing SGD for neural nets (with emphasis on Google s distributed neural net implementation)
More informationInception Network Overview. David White CS793
Inception Network Overview David White CS793 So, Leonardo DiCaprio dreams about dreaming... https://m.media-amazon.com/images/m/mv5bmjaxmzy3njcxnf5bml5banbnxkftztcwnti5otm0mw@@._v1_sy1000_cr0,0,675,1 000_AL_.jpg
More informationMultiple Kernel Learning for Emotion Recognition in the Wild
Multiple Kernel Learning for Emotion Recognition in the Wild Karan Sikka, Karmen Dykstra, Suchitra Sathyanarayana, Gwen Littlewort and Marian S. Bartlett Machine Perception Laboratory UCSD EmotiW Challenge,
More informationMachine Learning. The Breadth of ML Neural Networks & Deep Learning. Marc Toussaint. Duy Nguyen-Tuong. University of Stuttgart
Machine Learning The Breadth of ML Neural Networks & Deep Learning Marc Toussaint University of Stuttgart Duy Nguyen-Tuong Bosch Center for Artificial Intelligence Summer 2017 Neural Networks Consider
More informationBilevel Sparse Coding
Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional
More informationC. Poultney S. Cho pra (NYU Courant Institute) Y. LeCun
Efficient Learning of Sparse Overcomplete Representations with an Energy-Based Model Marc'Aurelio Ranzato C. Poultney S. Cho pra (NYU Courant Institute) Y. LeCun CIAR Summer School Toronto 2006 Why Extracting
More informationThe exam is closed book, closed notes except your one-page cheat sheet.
CS 189 Fall 2015 Introduction to Machine Learning Final Please do not turn over the page before you are instructed to do so. You have 2 hours and 50 minutes. Please write your initials on the top-right
More informationDeep Learning Cook Book
Deep Learning Cook Book Robert Haschke (CITEC) Overview Input Representation Output Layer + Cost Function Hidden Layer Units Initialization Regularization Input representation Choose an input representation
More informationStatistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte
Statistical Analysis of Metabolomics Data Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Outline Introduction Data pre-treatment 1. Normalization 2. Centering,
More informationNetwork Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Sources Hastie, Tibshirani, Friedman: The Elements of Statistical Learning James, Witten, Hastie, Tibshirani: An Introduction to Statistical Learning Andrew Ng:
More informationFace Recognition using SURF Features and SVM Classifier
International Journal of Electronics Engineering Research. ISSN 0975-6450 Volume 8, Number 1 (016) pp. 1-8 Research India Publications http://www.ripublication.com Face Recognition using SURF Features
More informationLecture 20: Neural Networks for NLP. Zubin Pahuja
Lecture 20: Neural Networks for NLP Zubin Pahuja zpahuja2@illinois.edu courses.engr.illinois.edu/cs447 CS447: Natural Language Processing 1 Today s Lecture Feed-forward neural networks as classifiers simple
More informationLEARNING A SPARSE DICTIONARY OF VIDEO STRUCTURE FOR ACTIVITY MODELING. Nandita M. Nayak, Amit K. Roy-Chowdhury. University of California, Riverside
LEARNING A SPARSE DICTIONARY OF VIDEO STRUCTURE FOR ACTIVITY MODELING Nandita M. Nayak, Amit K. Roy-Chowdhury University of California, Riverside ABSTRACT We present an approach which incorporates spatiotemporal
More informationDS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University
DS 4400 Machine Learning and Data Mining I Alina Oprea Associate Professor, CCIS Northeastern University January 24 2019 Logistics HW 1 is due on Friday 01/25 Project proposal: due Feb 21 1 page description
More informationNovel Lossy Compression Algorithms with Stacked Autoencoders
Novel Lossy Compression Algorithms with Stacked Autoencoders Anand Atreya and Daniel O Shea {aatreya, djoshea}@stanford.edu 11 December 2009 1. Introduction 1.1. Lossy compression Lossy compression is
More informationDeep Learning for Generic Object Recognition
Deep Learning for Generic Object Recognition, Computational and Biological Learning Lab The Courant Institute of Mathematical Sciences New York University Collaborators: Marc'Aurelio Ranzato, Fu Jie Huang,
More informationClassification of objects from Video Data (Group 30)
Classification of objects from Video Data (Group 30) Sheallika Singh 12665 Vibhuti Mahajan 12792 Aahitagni Mukherjee 12001 M Arvind 12385 1 Motivation Video surveillance has been employed for a long time
More informationApplying Supervised Learning
Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains
More informationSupervised vs unsupervised clustering
Classification Supervised vs unsupervised clustering Cluster analysis: Classes are not known a- priori. Classification: Classes are defined a-priori Sometimes called supervised clustering Extract useful
More informationon learned visual embedding patrick pérez Allegro Workshop Inria Rhônes-Alpes 22 July 2015
on learned visual embedding patrick pérez Allegro Workshop Inria Rhônes-Alpes 22 July 2015 Vector visual representation Fixed-size image representation High-dim (100 100,000) Generic, unsupervised: BoW,
More informationDeep Learning with Tensorflow AlexNet
Machine Learning and Computer Vision Group Deep Learning with Tensorflow http://cvml.ist.ac.at/courses/dlwt_w17/ AlexNet Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton, "Imagenet classification
More informationDEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla
DEEP LEARNING REVIEW Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature 2015 -Presented by Divya Chitimalla What is deep learning Deep learning allows computational models that are composed of multiple
More informationVulnerability of machine learning models to adversarial examples
Vulnerability of machine learning models to adversarial examples Petra Vidnerová Institute of Computer Science The Czech Academy of Sciences Hora Informaticae 1 Outline Introduction Works on adversarial
More informationA Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images
A Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images Marc Aurelio Ranzato Yann LeCun Courant Institute of Mathematical Sciences New York University - New York, NY 10003 Abstract
More informationFACIAL POINT DETECTION BASED ON A CONVOLUTIONAL NEURAL NETWORK WITH OPTIMAL MINI-BATCH PROCEDURE. Chubu University 1200, Matsumoto-cho, Kasugai, AICHI
FACIAL POINT DETECTION BASED ON A CONVOLUTIONAL NEURAL NETWORK WITH OPTIMAL MINI-BATCH PROCEDURE Masatoshi Kimura Takayoshi Yamashita Yu Yamauchi Hironobu Fuyoshi* Chubu University 1200, Matsumoto-cho,
More informationLarge-Scale Lasso and Elastic-Net Regularized Generalized Linear Models
Large-Scale Lasso and Elastic-Net Regularized Generalized Linear Models DB Tsai Steven Hillion Outline Introduction Linear / Nonlinear Classification Feature Engineering - Polynomial Expansion Big-data
More informationLearning Visual Semantics: Models, Massive Computation, and Innovative Applications
Learning Visual Semantics: Models, Massive Computation, and Innovative Applications Part II: Visual Features and Representations Liangliang Cao, IBM Watson Research Center Evolvement of Visual Features
More informationDeep Learning. Deep Learning provided breakthrough results in speech recognition and image classification. Why?
Data Mining Deep Learning Deep Learning provided breakthrough results in speech recognition and image classification. Why? Because Speech recognition and image classification are two basic examples of
More informationRobustness of Selective Desensitization Perceptron Against Irrelevant and Partially Relevant Features in Pattern Classification
Robustness of Selective Desensitization Perceptron Against Irrelevant and Partially Relevant Features in Pattern Classification Tomohiro Tanno, Kazumasa Horie, Jun Izawa, and Masahiko Morita University
More informationUnsupervised learning in Vision
Chapter 7 Unsupervised learning in Vision The fields of Computer Vision and Machine Learning complement each other in a very natural way: the aim of the former is to extract useful information from visual
More informationSingle Image Depth Estimation via Deep Learning
Single Image Depth Estimation via Deep Learning Wei Song Stanford University Stanford, CA Abstract The goal of the project is to apply direct supervised deep learning to the problem of monocular depth
More informationCOMP9444 Neural Networks and Deep Learning 5. Geometry of Hidden Units
COMP9 8s Geometry of Hidden Units COMP9 Neural Networks and Deep Learning 5. Geometry of Hidden Units Outline Geometry of Hidden Unit Activations Limitations of -layer networks Alternative transfer functions
More informationCAMCOS Report Day. December 9 th, 2015 San Jose State University Project Theme: Classification
CAMCOS Report Day December 9 th, 2015 San Jose State University Project Theme: Classification On Classification: An Empirical Study of Existing Algorithms based on two Kaggle Competitions Team 1 Team 2
More informationA Taxonomy of Semi-Supervised Learning Algorithms
A Taxonomy of Semi-Supervised Learning Algorithms Olivier Chapelle Max Planck Institute for Biological Cybernetics December 2005 Outline 1 Introduction 2 Generative models 3 Low density separation 4 Graph
More informationDoes the Brain do Inverse Graphics?
Does the Brain do Inverse Graphics? Geoffrey Hinton, Alex Krizhevsky, Navdeep Jaitly, Tijmen Tieleman & Yichuan Tang Department of Computer Science University of Toronto The representation used by the
More informationStatic Gesture Recognition with Restricted Boltzmann Machines
Static Gesture Recognition with Restricted Boltzmann Machines Peter O Donovan Department of Computer Science, University of Toronto 6 Kings College Rd, M5S 3G4, Canada odonovan@dgp.toronto.edu Abstract
More informationCost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling
[DOI: 10.2197/ipsjtcva.7.99] Express Paper Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling Takayoshi Yamashita 1,a) Takaya Nakamura 1 Hiroshi Fukui 1,b) Yuji
More informationClassifying Images with Visual/Textual Cues. By Steven Kappes and Yan Cao
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao Motivation Image search Building large sets of classified images Robotics Background Object recognition is unsolved Deformable shaped
More informationAutoencoders. Stephen Scott. Introduction. Basic Idea. Stacked AE. Denoising AE. Sparse AE. Contractive AE. Variational AE GAN.
Stacked Denoising Sparse Variational (Adapted from Paul Quint and Ian Goodfellow) Stacked Denoising Sparse Variational Autoencoding is training a network to replicate its input to its output Applications:
More informationPredict the Likelihood of Responding to Direct Mail Campaign in Consumer Lending Industry
Predict the Likelihood of Responding to Direct Mail Campaign in Consumer Lending Industry Jincheng Cao, SCPD Jincheng@stanford.edu 1. INTRODUCTION When running a direct mail campaign, it s common practice
More informationDoes the Brain do Inverse Graphics?
Does the Brain do Inverse Graphics? Geoffrey Hinton, Alex Krizhevsky, Navdeep Jaitly, Tijmen Tieleman & Yichuan Tang Department of Computer Science University of Toronto How to learn many layers of features
More informationISyE 6416 Basic Statistical Methods Spring 2016 Bonus Project: Big Data Report
ISyE 6416 Basic Statistical Methods Spring 2016 Bonus Project: Big Data Report Team Member Names: Caroline Roeger, Damon Frezza Project Title: Clustering and Classification of Handwritten Digits Responsibilities:
More informationSmart Content Recognition from Images Using a Mixture of Convolutional Neural Networks *
Smart Content Recognition from Images Using a Mixture of Convolutional Neural Networks * Tee Connie *, Mundher Al-Shabi *, and Michael Goh Faculty of Information Science and Technology, Multimedia University,
More informationPerceptron: This is convolution!
Perceptron: This is convolution! v v v Shared weights v Filter = local perceptron. Also called kernel. By pooling responses at different locations, we gain robustness to the exact spatial location of image
More informationAllstate Insurance Claims Severity: A Machine Learning Approach
Allstate Insurance Claims Severity: A Machine Learning Approach Rajeeva Gaur SUNet ID: rajeevag Jeff Pickelman SUNet ID: pattern Hongyi Wang SUNet ID: hongyiw I. INTRODUCTION The insurance industry has
More informationMachine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,
Machine Learning 10-701, Fall 2015 Deep Learning Eric Xing (and Pengtao Xie) Lecture 8, October 6, 2015 Eric Xing @ CMU, 2015 1 A perennial challenge in computer vision: feature engineering SIFT Spin image
More informationMACHINE LEARNING: CLUSTERING, AND CLASSIFICATION. Steve Tjoa June 25, 2014
MACHINE LEARNING: CLUSTERING, AND CLASSIFICATION Steve Tjoa kiemyang@gmail.com June 25, 2014 Review from Day 2 Supervised vs. Unsupervised Unsupervised - clustering Supervised binary classifiers (2 classes)
More informationFacial Expression Recognition Using Non-negative Matrix Factorization
Facial Expression Recognition Using Non-negative Matrix Factorization Symeon Nikitidis, Anastasios Tefas and Ioannis Pitas Artificial Intelligence & Information Analysis Lab Department of Informatics Aristotle,
More informationUsing Machine Learning to Identify Security Issues in Open-Source Libraries. Asankhaya Sharma Yaqin Zhou SourceClear
Using Machine Learning to Identify Security Issues in Open-Source Libraries Asankhaya Sharma Yaqin Zhou SourceClear Outline - Overview of problem space Unidentified security issues How Machine Learning
More informationA Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images
A Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images Marc Aurelio Ranzato Yann LeCun Courant Institute of Mathematical Sciences New York University - New York, NY 10003 Abstract
More informationDeep Learning. Volker Tresp Summer 2014
Deep Learning Volker Tresp Summer 2014 1 Neural Network Winter and Revival While Machine Learning was flourishing, there was a Neural Network winter (late 1990 s until late 2000 s) Around 2010 there
More informationDeep Convolutional Neural Network using Triplet of Faces, Deep Ensemble, and Scorelevel Fusion for Face Recognition
IEEE 2017 Conference on Computer Vision and Pattern Recognition Deep Convolutional Neural Network using Triplet of Faces, Deep Ensemble, and Scorelevel Fusion for Face Recognition Bong-Nam Kang*, Yonghyun
More informationFacial Expression Recognition Using a Hybrid CNN SIFT Aggregator
Facial Expression Recognition Using a Hybrid CNN SIFT Aggregator Mundher Al-Shabi, Wooi Ping Cheah, Tee Connie Faculty of Information Science and Technology, Multimedia University, Melaka, Malaysia Abstract.
More informationCPSC 340: Machine Learning and Data Mining. Deep Learning Fall 2018
CPSC 340: Machine Learning and Data Mining Deep Learning Fall 2018 Last Time: Multi-Dimensional Scaling Multi-dimensional scaling (MDS): Non-parametric visualization: directly optimize the z i locations.
More informationAn Analysis of Single-Layer Networks in Unsupervised Feature Learning
An Analysis of Single-Layer Networks in Unsupervised Feature Learning Adam Coates Honglak Lee Andrew Y. Ng Stanford University Computer Science Dept. 353 Serra Mall Stanford, CA 94305 University of Michigan
More informationarxiv: v1 [cs.lg] 20 Dec 2013
Unsupervised Feature Learning by Deep Sparse Coding Yunlong He Koray Kavukcuoglu Yun Wang Arthur Szlam Yanjun Qi arxiv:1312.5783v1 [cs.lg] 20 Dec 2013 Abstract In this paper, we propose a new unsupervised
More informationMachine Learning in Python. Rohith Mohan GradQuant Spring 2018
Machine Learning in Python Rohith Mohan GradQuant Spring 2018 What is Machine Learning? https://twitter.com/myusuf3/status/995425049170489344 Traditional Programming Data Computer Program Output Getting
More informationCIS581: Computer Vision and Computational Photography Project 4, Part B: Convolutional Neural Networks (CNNs) Due: Dec.11, 2017 at 11:59 pm
CIS581: Computer Vision and Computational Photography Project 4, Part B: Convolutional Neural Networks (CNNs) Due: Dec.11, 2017 at 11:59 pm Instructions CNNs is a team project. The maximum size of a team
More informationMulti-Task Self-Supervised Visual Learning
Multi-Task Self-Supervised Visual Learning Sikai Zhong March 4, 2018 COMPUTER SCIENCE Table of contents 1. Introduction 2. Self-supervised Tasks 3. Architectures 4. Experiments 1 Introduction Self-supervised
More informationIntroduction to Data Science. Introduction to Data Science with Python. Python Basics: Basic Syntax, Data Structures. Python Concepts (Core)
Introduction to Data Science What is Analytics and Data Science? Overview of Data Science and Analytics Why Analytics is is becoming popular now? Application of Analytics in business Analytics Vs Data
More informationUnsupervised Learning of Spatiotemporally Coherent Metrics
Unsupervised Learning of Spatiotemporally Coherent Metrics Ross Goroshin, Joan Bruna, Jonathan Tompson, David Eigen, Yann LeCun arxiv 2015. Presented by Jackie Chu Contributions Insight between slow feature
More informationImage enhancement for face recognition using color segmentation and Edge detection algorithm
Image enhancement for face recognition using color segmentation and Edge detection algorithm 1 Dr. K Perumal and 2 N Saravana Perumal 1 Computer Centre, Madurai Kamaraj University, Madurai-625021, Tamilnadu,
More informationFinal Report: Kaggle Soil Property Prediction Challenge
Final Report: Kaggle Soil Property Prediction Challenge Saurabh Verma (verma076@umn.edu, (612)598-1893) 1 Project Goal Low cost and rapid analysis of soil samples using infrared spectroscopy provide new
More informationCPSC 340: Machine Learning and Data Mining. Multi-Dimensional Scaling Fall 2017
CPSC 340: Machine Learning and Data Mining Multi-Dimensional Scaling Fall 2017 Assignment 4: Admin 1 late day for tonight, 2 late days for Wednesday. Assignment 5: Due Monday of next week. Final: Details
More informationThe exam is closed book, closed notes except your one-page (two-sided) cheat sheet.
CS 189 Spring 2015 Introduction to Machine Learning Final You have 2 hours 50 minutes for the exam. The exam is closed book, closed notes except your one-page (two-sided) cheat sheet. No calculators or
More information10/14/2017. Dejan Sarka. Anomaly Detection. Sponsors
Dejan Sarka Anomaly Detection Sponsors About me SQL Server MVP (17 years) and MCT (20 years) 25 years working with SQL Server Authoring 16 th book Authoring many courses, articles Agenda Introduction Simple
More informationLogistic Regression
Logistic Regression ddebarr@uw.edu 2016-05-26 Agenda Model Specification Model Fitting Bayesian Logistic Regression Online Learning and Stochastic Optimization Generative versus Discriminative Classifiers
More informationRyerson University CP8208. Soft Computing and Machine Intelligence. Naive Road-Detection using CNNS. Authors: Sarah Asiri - Domenic Curro
Ryerson University CP8208 Soft Computing and Machine Intelligence Naive Road-Detection using CNNS Authors: Sarah Asiri - Domenic Curro April 24 2016 Contents 1 Abstract 2 2 Introduction 2 3 Motivation
More informationBeyond Bags of Features
: for Recognizing Natural Scene Categories Matching and Modeling Seminar Instructed by Prof. Haim J. Wolfson School of Computer Science Tel Aviv University December 9 th, 2015
More informationDeep Learning With Noise
Deep Learning With Noise Yixin Luo Computer Science Department Carnegie Mellon University yixinluo@cs.cmu.edu Fan Yang Department of Mathematical Sciences Carnegie Mellon University fanyang1@andrew.cmu.edu
More informationIMPROVED FACE RECOGNITION USING ICP TECHNIQUES INCAMERA SURVEILLANCE SYSTEMS. Kirthiga, M.E-Communication system, PREC, Thanjavur
IMPROVED FACE RECOGNITION USING ICP TECHNIQUES INCAMERA SURVEILLANCE SYSTEMS Kirthiga, M.E-Communication system, PREC, Thanjavur R.Kannan,Assistant professor,prec Abstract: Face Recognition is important
More information