Object Recognition Under Complex Environmental Conditions on Remote Sensing Image
|
|
- Jared Warner
- 5 years ago
- Views:
Transcription
1 International Journal of Applied Environmental Sciences ISSN Volume 11, Number 3 (016), pp Research India Publications Object Recognition Under Complex Environmental Conditions on Remote Sensing Image Jian-min Liu* 1, and Min-hua Yang 1 1 School of Geosciences and Info-Physics, Central South University, Changsha, 41000, China. School of Information Science and Engineering, Hunan Institute of Humanities Science and echnology, Loudi , China. Abstract Object detection on high-resolution and optical remote sensing images has attracted extensive interest. Because Deep neural networks have boomed in recent years, based on caffe and cuda, We trained a multilayer convolutional neural network to classify the 1 thousand high-resolution and unlabeled optical remote sensing images via the Internet into the 1000 different classes. On the test dataset, we achieved error rates of 17.7% which work went relatively well than the previous machine learning techniques. o help to save training time, we also used pre-trained model, GX970 GPU and powerful implementation of the algorithm, these approach helping we saving a lot of time. he results compared with the published ones, and good agreement is obtained. Keywords: Deep Neural Network(DNN); Remote Sensing Image; Detection 1. Introduction In military techniques and many other fields, aircraft detection is increasingly used. Various types classification methods have been applied to remote sensing work as object classifiers. Conventional remote sensing image detection techniques mainly based on spectral, and the detection algorithms typically include linear regression, nonlinear regression, logistic regression, k-means, conditional random fields and maximumlikelihood[1][]. hose like above can be divided into the method is based on statistical methods and non statistical methods. Firstly, statistical based method y x; ~ (, ) can be divided into regression problem and classification problem y x; ~ Bernoulli( ).μandφthat is said that x 和, both of them look different, but in
2 75 Jian-min Liu and Min-hua Yang fact, are two special cases of general linear model with a model that Michael.I.Jordan proposed. Mathematical expression can be written as: p( y; ) b( y)exp( ( y) a ( )) (1) Extending from logistic regression method, it is very common that classification among variety objects except two classification method in reality, such as remote sensing image target for farmland, soil, water, road...,let y {1,,..., k}, then p( y; ) exp(((y)) 1 log( 1 / k) ( ( y)) log( / k) ()... ( ( y)) log( / ) l og( ) let 1 k1 k1 k k log( / k ) log( / ) k, a( ) log( k ),b(y) 1 log( k 1 / k) hen (1) is expressed as: p( y; ) b( y)exp( ( y) a( )), and it is called softmax algorithm. On the one hand softmax algorithm can be understood as a generalization of the logistic algorithm, but on the other hand the logistic algorithm is a special case of the softmax algorithm. he above methods is based on the statistical analysis of the spectral and spatial information of each pixel. Secondly, non statistical methods. based method typically include Neural network and Support vector machine algorithm, etc. In the face and handwritten numeral recognition, remote sensing image classification and other applications, it is one of the best machine learning methods before deep learning boomed. he researchers made a lot of results by using spectral information, spatial information and object information. On many occasions, SVM-based classifiers work went relatively well than than other extensive used machine learning techniques[][3]. On the MNIS dataset handwritten digit recognition task, the highest-accuracy reports show backpropagation hit 1.6% of error rates only if the initial weights has been initialized randomly and support vector machines hit 1.4% of error rates[3]. Owing to the complexity a long time, these classifiers has been locked in the state-of-theart methods[4][5]. Given these issues, we consider the challenge of character and digital recognition, aircraft detection and a gradually improving understanding of the human's feelings[6]. In years gone by the expectant, based on hand-engineering approach these issues are certainly laborintensive, and work inefficient for new problems. Some nonsupervised feature representations ones was expected to replace hand-engineered methods[4][6]. DNN algorithm like sparse autoencoder, deep boltzmann machines and deep convolutional neural network method equip scientist and R&D engineers to classify, detect and recognize various types objects. A large, deep neural network, such as sparse autoencoder nonsupervised learn features from high-dimensional original data, and train central layer which consist of a lot of hidden units to reconstruct original input vectors, and detect edges at key positions and structure of the image[4][6][7]. o judge the weights of each unit of each layer, gradient descent is susceptible to local fine-tuning in such a autoencoder network. Speaking from experience gradient descent works well only if the initial weights has been initialized outline low-level structures such as edges from it, based on this, and form a just and comprehensible view of high-level structures such as corners, structure, and outline. he feature acquired from unsupervised learning can be used to reconstruct original input (3)
3 Object Recognition Under Complex Environmental Conditions 753 vectors[8]. After fine-tuning, a network with three hidden layers forms a very good generative model of the joint distribution of handwritten digit images and their labels. his generative model gives better digit classification than the best discriminative learning algorithms[9]. Speech recognition is a classic application of neural networks, in recent years, and deep learning has achieved remarkable results in speech recognition in recent years. Microsoft and Google have been gradually put deep learning's speech recognition algorithm into commercial applications. Microsoft's original speech recognition based on the use of hybrid algorithm Gauss reduce the error rate of words to 16%. Bengio, Hinton and other scholars solved MNIS based classification problems, and breaking the old mark 0.14% error rate Set by traditional machine learning algorithm for many years[10][11]. Krizhevsky had made significant progress on the ImageNet data set with a 15.3% error rate[1]. he performance of traditional machine learning methods rely heavily on the choice skills representation or feature method of object data, which can be called an art of scientific research rather than a technology, need to be trained with regularity by researchers and is extremely strict by application environment. So the actual effect of traditional machine learning algorithms depend on pretreatment technology and data conversion which whether can support the machine learning algorithm. he process of selecting such characteristics is important but time-consuming, and algorithms can extract and organize data from independent primary characteristics. he implementation of production process characteristics are selected from massive data based on human intelligence of researcher's decision instead of training with regularity unsupervised operation. Human intelligence can make up for the weakness of the algorithm. In order to expand the application field of the machine learning algorithm, reduce the labor cost, we must reduce the algorithm of feature selection dependence on trained researcher, and realize true artificial intelligence, automation and intelligent independent from the brain[13]. his paper presents aircraft detection on high resolution optical remote sensing images based on Deep Neural Network which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax. he results are compared with the published ones. Finaly, because GPU's memory available is scant, the dimension of network is limited to a certain location. Our network spends about 19 hours to train on one GX 750i GB GPUs.. HEOREICAL ANALYSIS A common Deep neural network consists of autoencoder layer, convolutional layer, pooling layer, and fully-connected layer. Similar to the conventional method, and for the sake of convenience, the equations of overall cost of backpropagation is: n mu - 1 sl s u+1 ( i ) ( i ) ( u) V, d i j j =1 u=1 j =1 i =1 1 1 γ C( V, d) =( k ( x - y ) ) ) + ( V ) n 4 compute partial derivation, and expressed as (4)
4 754 Jian-min Liu and Min-hua Yang V n 1 γ CV (, d) ( CV (, d; x, y ) ) V n C ( u) ( u) j i j 1 j i We obtain: CV (, d; x, y ) ( u) V ji CV (, d; x, y ) ( u) d ( j ) ( j ) ( u) ji a ( u) ( u1) i j ( u 1) j i (7) After that: C sp 1 (, ) s C V d log (1 )log ^ ^ i1 1 1 (( ( )) ( ) () () (3) ' s j Vij 0 g z ^ ^ i1 i 1 i i i (9) After that, the convolutional layer execute an image convolution for output of the previous layer, where the filter of convolution layer named learned core which consists 16x16 matrix of feature. he layer after fully-connected layer works with a final multi-way softmax. On this layer, we choose softmax model which generalizes logistic model in multi-class classification and recognition. ( j) ( j) p( y 1 x ; w) ( j) ( j) p( y x ; w) ( j) ( j) p( y 3 x ; w) ( j) kw( x )... ( j) ( j) p( y h x ; w) (10) Because summary of probability is equal to 1, we obtain w1 e w e w3 e ( j) 1 k ( ). w x h wi e. i1. wh e (11) We obtain a working model of softmax which come into play in multi-class classification and recognition. (5) (6) (8)
5 Object Recognition Under Complex Environmental Conditions Experimental results and analysis: In this Experiment, We focus on aircraft detection on high-resolution and optical remote sensing images. Based on caffe and cuda, we trained a multilayer convolutional neural network to classify the 1 thousand high-resolution and unlabeled optical remote sensing images via the Internet into the 1000 different classes. On the test dataset, we achieved error rates of 17.7% which work went relatively well than the previous machine learning techniques.. o help to save training time, we also used pre-trained model, GX970 GPU and powerful implementation of the algorithm, these approach helping we saving a lot of time. he process from the first layer to the seventh layer of the classification of an image is shown in Figure 1-(a-l). Figure 1-(a) show the input raw image, Figure 1-(b) show the output of first layer named conv1 layer, Figure 1-(c) show the filters of second layer named conv layer, Figure 1-(d) show the output of second layer named conv layer, Figure 1-(e) show the output of third layer output named conv3 layer, Figure 1- (f) show the output of fourth layer named conv4 layer, Figure 1-(g) show the output of fifth layer named conv5 layer, Figure 1-(h) show the fifth layer after pooling named pool5 layer, Figure 1-(i) show the first fully connected layer named fc6 layer, Figure 1-(j) show the second fully connected layer named fc7 layer, Figure 1-(k) show the output of final probability, and Figure 1-(l) show the top 5 detected labels. he top 1 detected label is space shuttle, the top detected label is aircraft carrier, the top 3 detected label is liner, ocean liner, the top 4 detected label: speedboat, and the top 5 detected label is yawl.
6 756 Jian-min Liu and Min-hua Yang Figure 1: Experimental results. (a) he input image (b)he first layer output, conv1 (c) he second layer filters, conv (d) he second layer output, conv (e) he third layer output, conv3 (f) he fourth layer output, conv4 (g) he fifth layer output, conv5 (h) he fifth layer after pooling, pool5 (i) he first fully connected layer, fc6 (j) he second fully connected layer, fc7 (k) he final probability output (l) he top 5 predicted labels. 4. Conclusion his paper on the one hand provide an extension to the existing analytical methods of aircraft detection on high-resolution and optical remote sensing images, on the other hand, it maybe help to further the development of this Deep Neural Network means. Detail See able 1. able1: Comparision of Convolutional Neural Network and traditional method Convolutional Neural Network raditional Method Complexity A highly complex multilayer Overall structure is not convolutional neural network is complicated, but Experienced constructed manual intervention is researchers and a lot of time are unnecessary needed Input Requirment Original unlabeled image Pre processed image Memory Sharply increase the memory Less memory requirement, but requirement requirement of training time more manual intervention. Does not deduce the memory requirement of working time Simulation Reduce CPU and GPU time Long CPU time is required time siginificantly. GPU not Available About from 1/3 to 1/10 time of that of traditional method. Acknowledgments Supported by Provincial Natural Science Foundation of Hunan ( 01JJ509 ) Supported by Science and echnology Planning Project of Loudi (014KJ04)
7 Object Recognition Under Complex Environmental Conditions 757 References [1] G. M. Foody and A. Mathur, A relative evaluation of multiclass image classification by support vector machines[j], IEEE rans. Geosci. Remote Sens., vol. 4, no. 6, pp , Jun [] Yushi Chen. Deep Learning-Based Classification of Hyperspectral Data[J], IEEE JOURNAL OF SELECED OPICS IN APPLIED EARH OBSERVAIONS AND REMOE SENSING, VOL. 7, NO. 6, JUNE [3] J. A. Gualtieri and S. Chettri. Support vector machines for classification of hyperspectral data[p], Proc. IEEE Geosci. Remote Sens. Symp. (IGARSS),Honolulu, HI, USA, 000, pp [4] Honglak Lee Chaitanya Ekanadham Andrew Y. Ng. Sparse deep belief net model for visual area V[P].NIPS 007. [5] L. Zhuo et al., Agenetic algorithm based wrapper feature selection method for classification of hyperspectral images using support vector machine[p], Proc. Geoinformat. Joint Conf. GIS Built Environ. Classif. Remote Sens. Images Int. Soc. Opt. Photonics, Nov. 008, pp J 71471J.]. [6] Honglak Lee, Roger Grosse, Rajesh Ranganath and Andrew Y. Ng. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations, ICML 009. [7] G. E. Hinton* and R. R. Salakhutdinov Reducing the Dimensionality of Data with Neural Networks SCIENCE VOL JULY 006 [8] Honglak Lee, Roger Grosse, Rajesh Ranganath, and Andrew Y. Ng. Unsupervised Learning of Hierarchical Representations with Convolutional Deep Belief Networks.Communications of HE ACM VOL.54 NO.10, [9] Geoffrey E. Hinton A Fast Learning Algorithm for Deep Belief Nets.Neural Computation 18, (006)] [10] G. E. Hinton,S. Osindero,Y. W. eh.a fast learning algorithm for deep belief nets. Neural Computation. 006 [11] Y. Bengio, G. Mesnil, Y. Dauphin, and S. Rifai, Better Mixing via Deep Representations, Proc. Int l Conf. Machine Learning, 013. [1] Krizhevsky, A., Sutskever, I. and Hinton, G. E. ImageNet Classification with Deep Convolutional Neural Networks. NIPS 01: Neural Information Processing Systems, Lake ahoe, Nevada [13] Liu Jianmin, Huang Fan, Dai Jun. Image recognition technology for remote sensing based on research and application Machine learning. Journal of Xi'an University of Arts & Science. Oct. 015
8 758 Jian-min Liu and Min-hua Yang Authors LIU Jian-min, PhD student Research interests: Remote Sensing, GIS, and Machine learning,
ImageNet Classification with Deep Convolutional Neural Networks
ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky Ilya Sutskever Geoffrey Hinton University of Toronto Canada Paper with same name to appear in NIPS 2012 Main idea Architecture
More informationDeep Learning Workshop. Nov. 20, 2015 Andrew Fishberg, Rowan Zellers
Deep Learning Workshop Nov. 20, 2015 Andrew Fishberg, Rowan Zellers Why deep learning? The ImageNet Challenge Goal: image classification with 1000 categories Top 5 error rate of 15%. Krizhevsky, Alex,
More informationDEEP LEARNING TO DIVERSIFY BELIEF NETWORKS FOR REMOTE SENSING IMAGE CLASSIFICATION
DEEP LEARNING TO DIVERSIFY BELIEF NETWORKS FOR REMOTE SENSING IMAGE CLASSIFICATION S.Dhanalakshmi #1 #PG Scholar, Department of Computer Science, Dr.Sivanthi Aditanar college of Engineering, Tiruchendur
More informationDeep Learning for Computer Vision II
IIIT Hyderabad Deep Learning for Computer Vision II C. V. Jawahar Paradigm Shift Feature Extraction (SIFT, HoG, ) Part Models / Encoding Classifier Sparrow Feature Learning Classifier Sparrow L 1 L 2 L
More informationDeep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks
Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Si Chen The George Washington University sichen@gwmail.gwu.edu Meera Hahn Emory University mhahn7@emory.edu Mentor: Afshin
More informationMachine Learning. MGS Lecture 3: Deep Learning
Dr Michel F. Valstar http://cs.nott.ac.uk/~mfv/ Machine Learning MGS Lecture 3: Deep Learning Dr Michel F. Valstar http://cs.nott.ac.uk/~mfv/ WHAT IS DEEP LEARNING? Shallow network: Only one hidden layer
More informationCS 2750: Machine Learning. Neural Networks. Prof. Adriana Kovashka University of Pittsburgh April 13, 2016
CS 2750: Machine Learning Neural Networks Prof. Adriana Kovashka University of Pittsburgh April 13, 2016 Plan for today Neural network definition and examples Training neural networks (backprop) Convolutional
More informationIntroduction to Deep Learning
ENEE698A : Machine Learning Seminar Introduction to Deep Learning Raviteja Vemulapalli Image credit: [LeCun 1998] Resources Unsupervised feature learning and deep learning (UFLDL) tutorial (http://ufldl.stanford.edu/wiki/index.php/ufldl_tutorial)
More informationProceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong
, March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong TABLE I CLASSIFICATION ACCURACY OF DIFFERENT PRE-TRAINED MODELS ON THE TEST DATA
More informationConvolutional Neural Networks
Lecturer: Barnabas Poczos Introduction to Machine Learning (Lecture Notes) Convolutional Neural Networks Disclaimer: These notes have not been subjected to the usual scrutiny reserved for formal publications.
More informationUnsupervised Deep Learning for Scene Recognition
Unsupervised Deep Learning for Scene Recognition Akram Helou and Chau Nguyen May 19, 2011 1 Introduction Object and scene recognition are usually studied separately. However, research [2]shows that context
More informationObject Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal
Object Detection Lecture 10.3 - Introduction to deep learning (CNN) Idar Dyrdal Deep Learning Labels Computational models composed of multiple processing layers (non-linear transformations) Used to learn
More informationDeep Learning & Neural Networks
Deep Learning & Neural Networks Machine Learning CSE4546 Sham Kakade University of Washington November 29, 2016 Sham Kakade 1 Announcements: HW4 posted Poster Session Thurs, Dec 8 Today: Review: EM Neural
More informationMachine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,
Machine Learning 10-701, Fall 2015 Deep Learning Eric Xing (and Pengtao Xie) Lecture 8, October 6, 2015 Eric Xing @ CMU, 2015 1 A perennial challenge in computer vision: feature engineering SIFT Spin image
More informationNeural Networks. Single-layer neural network. CSE 446: Machine Learning Emily Fox University of Washington March 10, /10/2017
3/0/207 Neural Networks Emily Fox University of Washington March 0, 207 Slides adapted from Ali Farhadi (via Carlos Guestrin and Luke Zettlemoyer) Single-layer neural network 3/0/207 Perceptron as a neural
More informationNeural Networks for Machine Learning. Lecture 15a From Principal Components Analysis to Autoencoders
Neural Networks for Machine Learning Lecture 15a From Principal Components Analysis to Autoencoders Geoffrey Hinton Nitish Srivastava, Kevin Swersky Tijmen Tieleman Abdel-rahman Mohamed Principal Components
More informationPouya Kousha Fall 2018 CSE 5194 Prof. DK Panda
Pouya Kousha Fall 2018 CSE 5194 Prof. DK Panda 1 Observe novel applicability of DL techniques in Big Data Analytics. Applications of DL techniques for common Big Data Analytics problems. Semantic indexing
More informationDeep Learning With Noise
Deep Learning With Noise Yixin Luo Computer Science Department Carnegie Mellon University yixinluo@cs.cmu.edu Fan Yang Department of Mathematical Sciences Carnegie Mellon University fanyang1@andrew.cmu.edu
More informationDeep Learning and Its Applications
Convolutional Neural Network and Its Application in Image Recognition Oct 28, 2016 Outline 1 A Motivating Example 2 The Convolutional Neural Network (CNN) Model 3 Training the CNN Model 4 Issues and Recent
More informationGated Boltzmann Machine in Texture Modeling
Gated Boltzmann Machine in Texture Modeling Tele Hao, Tapani Rao, Alexander Ilin, and Juha Karhunen Department of Information and Computer Science Aalto University, Espoo, Finland firstname.lastname@aalto.fi
More informationNeural Networks for unsupervised learning From Principal Components Analysis to Autoencoders to semantic hashing
Neural Networks for unsupervised learning From Principal Components Analysis to Autoencoders to semantic hashing feature 3 PC 3 Beate Sick Many slides are taken form Hinton s great lecture on NN: https://www.coursera.org/course/neuralnets
More informationMachine Learning 13. week
Machine Learning 13. week Deep Learning Convolutional Neural Network Recurrent Neural Network 1 Why Deep Learning is so Popular? 1. Increase in the amount of data Thanks to the Internet, huge amount of
More informationMachine Learning. The Breadth of ML Neural Networks & Deep Learning. Marc Toussaint. Duy Nguyen-Tuong. University of Stuttgart
Machine Learning The Breadth of ML Neural Networks & Deep Learning Marc Toussaint University of Stuttgart Duy Nguyen-Tuong Bosch Center for Artificial Intelligence Summer 2017 Neural Networks Consider
More informationDeep Learning with Tensorflow AlexNet
Machine Learning and Computer Vision Group Deep Learning with Tensorflow http://cvml.ist.ac.at/courses/dlwt_w17/ AlexNet Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton, "Imagenet classification
More informationImplementation of Deep Convolutional Neural Net on a Digital Signal Processor
Implementation of Deep Convolutional Neural Net on a Digital Signal Processor Elaina Chai December 12, 2014 1. Abstract In this paper I will discuss the feasibility of an implementation of an algorithm
More informationConvolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech
Convolutional Neural Networks Computer Vision Jia-Bin Huang, Virginia Tech Today s class Overview Convolutional Neural Network (CNN) Training CNN Understanding and Visualizing CNN Image Categorization:
More informationTo be Bernoulli or to be Gaussian, for a Restricted Boltzmann Machine
2014 22nd International Conference on Pattern Recognition To be Bernoulli or to be Gaussian, for a Restricted Boltzmann Machine Takayoshi Yamashita, Masayuki Tanaka, Eiji Yoshida, Yuji Yamauchi and Hironobu
More informationNeural Networks and Deep Learning
Neural Networks and Deep Learning Example Learning Problem Example Learning Problem Celebrity Faces in the Wild Machine Learning Pipeline Raw data Feature extract. Feature computation Inference: prediction,
More informationMulti-Glance Attention Models For Image Classification
Multi-Glance Attention Models For Image Classification Chinmay Duvedi Stanford University Stanford, CA cduvedi@stanford.edu Pararth Shah Stanford University Stanford, CA pararth@stanford.edu Abstract We
More informationA Deep Learning Framework for Authorship Classification of Paintings
A Deep Learning Framework for Authorship Classification of Paintings Kai-Lung Hua ( 花凱龍 ) Dept. of Computer Science and Information Engineering National Taiwan University of Science and Technology Taipei,
More informationA Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images
A Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images Marc Aurelio Ranzato Yann LeCun Courant Institute of Mathematical Sciences New York University - New York, NY 10003 Abstract
More informationApparel Classifier and Recommender using Deep Learning
Apparel Classifier and Recommender using Deep Learning Live Demo at: http://saurabhg.me/projects/tag-that-apparel Saurabh Gupta sag043@ucsd.edu Siddhartha Agarwal siagarwa@ucsd.edu Apoorve Dave a1dave@ucsd.edu
More informationImage Classification using Fast Learning Convolutional Neural Networks
, pp.50-55 http://dx.doi.org/10.14257/astl.2015.113.11 Image Classification using Fast Learning Convolutional Neural Networks Keonhee Lee 1 and Dong-Chul Park 2 1 Software Device Research Center Korea
More informationLarge Scale Distributed Deep Networks
Large Scale Distributed Deep Networks Yifu Huang School of Computer Science, Fudan University huangyifu@fudan.edu.cn COMP630030 Data Intensive Computing Report, 2013 Yifu Huang (FDU CS) COMP630030 Report
More informationDeep Learning Basic Lecture - Complex Systems & Artificial Intelligence 2017/18 (VO) Asan Agibetov, PhD.
Deep Learning 861.061 Basic Lecture - Complex Systems & Artificial Intelligence 2017/18 (VO) Asan Agibetov, PhD asan.agibetov@meduniwien.ac.at Medical University of Vienna Center for Medical Statistics,
More informationA Deep Learning primer
A Deep Learning primer Riccardo Zanella r.zanella@cineca.it SuperComputing Applications and Innovation Department 1/21 Table of Contents Deep Learning: a review Representation Learning methods DL Applications
More informationNovel Lossy Compression Algorithms with Stacked Autoencoders
Novel Lossy Compression Algorithms with Stacked Autoencoders Anand Atreya and Daniel O Shea {aatreya, djoshea}@stanford.edu 11 December 2009 1. Introduction 1.1. Lossy compression Lossy compression is
More informationDeep Neural Networks:
Deep Neural Networks: Part II Convolutional Neural Network (CNN) Yuan-Kai Wang, 2016 Web site of this course: http://pattern-recognition.weebly.com source: CNN for ImageClassification, by S. Lazebnik,
More informationA FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen
A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS Kuan-Chuan Peng and Tsuhan Chen School of Electrical and Computer Engineering, Cornell University, Ithaca, NY
More informationClassifying Depositional Environments in Satellite Images
Classifying Depositional Environments in Satellite Images Alex Miltenberger and Rayan Kanfar Department of Geophysics School of Earth, Energy, and Environmental Sciences Stanford University 1 Introduction
More informationPoS(CENet2017)005. The Classification of Hyperspectral Images Based on Band-Grouping and Convolutional Neural Network. Speaker.
The Classification of Hyperspectral Images Based on Band-Grouping and Convolutional Neural Network 1 Xi an Hi-Tech Institute Xi an 710025, China E-mail: dr-f@21cnl.c Hongyang Gu Xi an Hi-Tech Institute
More informationNVIDIA FOR DEEP LEARNING. Bill Veenhuis
NVIDIA FOR DEEP LEARNING Bill Veenhuis bveenhuis@nvidia.com Nvidia is the world s leading ai platform ONE ARCHITECTURE CUDA 2 GPU: Perfect Companion for Accelerating Apps & A.I. CPU GPU 3 Intro to AI AGENDA
More informationCMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro
CMU 15-781 Lecture 18: Deep learning and Vision: Convolutional neural networks Teacher: Gianni A. Di Caro DEEP, SHALLOW, CONNECTED, SPARSE? Fully connected multi-layer feed-forward perceptrons: More powerful
More informationCONVOLUTIONAL NEURAL NETWORKS FOR HIGH SPATIAL RESOLUTION REMOTE SENSING IMAGE CLASSIFICATION
CONVOLUTIONAL NEURAL NETWORKS FOR HIGH SPATIAL RESOLUTION REMOTE SENSING IMAGE CLASSIFICATION 1. Sakhi. G 1, 2., R. Balasubramanian 2, R.Nedunchezhian 3 1 Research Scholar, Manonmaniam Sundaranar University,Tirunelveli
More informationNeural Network Neurons
Neural Networks Neural Network Neurons 1 Receives n inputs (plus a bias term) Multiplies each input by its weight Applies activation function to the sum of results Outputs result Activation Functions Given
More informationEND-TO-END CHINESE TEXT RECOGNITION
END-TO-END CHINESE TEXT RECOGNITION Jie Hu 1, Tszhang Guo 1, Ji Cao 2, Changshui Zhang 1 1 Department of Automation, Tsinghua University 2 Beijing SinoVoice Technology November 15, 2017 Presentation at
More informationAutoencoders, denoising autoencoders, and learning deep networks
4 th CiFAR Summer School on Learning and Vision in Biology and Engineering Toronto, August 5-9 2008 Autoencoders, denoising autoencoders, and learning deep networks Part II joint work with Hugo Larochelle,
More informationLecture 19: Generative Adversarial Networks
Lecture 19: Generative Adversarial Networks Roger Grosse 1 Introduction Generative modeling is a type of machine learning where the aim is to model the distribution that a given set of data (e.g. images,
More informationSEMANTIC COMPUTING. Lecture 8: Introduction to Deep Learning. TU Dresden, 7 December Dagmar Gromann International Center For Computational Logic
SEMANTIC COMPUTING Lecture 8: Introduction to Deep Learning Dagmar Gromann International Center For Computational Logic TU Dresden, 7 December 2018 Overview Introduction Deep Learning General Neural Networks
More informationHENet: A Highly Efficient Convolutional Neural. Networks Optimized for Accuracy, Speed and Storage
HENet: A Highly Efficient Convolutional Neural Networks Optimized for Accuracy, Speed and Storage Qiuyu Zhu Shanghai University zhuqiuyu@staff.shu.edu.cn Ruixin Zhang Shanghai University chriszhang96@shu.edu.cn
More informationHYPERSPECTRAL imagery has been increasingly used
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 14, NO. 5, MAY 2017 597 Transferred Deep Learning for Anomaly Detection in Hyperspectral Imagery Wei Li, Senior Member, IEEE, Guodong Wu, and Qian Du, Senior
More informationStudy of Residual Networks for Image Recognition
Study of Residual Networks for Image Recognition Mohammad Sadegh Ebrahimi Stanford University sadegh@stanford.edu Hossein Karkeh Abadi Stanford University hosseink@stanford.edu Abstract Deep neural networks
More information3D Object Recognition with Deep Belief Nets
3D Object Recognition with Deep Belief Nets Vinod Nair and Geoffrey E. Hinton Department of Computer Science, University of Toronto 10 King s College Road, Toronto, M5S 3G5 Canada {vnair,hinton}@cs.toronto.edu
More informationHand Written Digit Recognition Using Tensorflow and Python
Hand Written Digit Recognition Using Tensorflow and Python Shekhar Shiroor Department of Computer Science College of Engineering and Computer Science California State University-Sacramento Sacramento,
More informationFace Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN
2016 International Conference on Artificial Intelligence: Techniques and Applications (AITA 2016) ISBN: 978-1-60595-389-2 Face Recognition Using Vector Quantization Histogram and Support Vector Machine
More informationA Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images
A Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images Marc Aurelio Ranzato Yann LeCun Courant Institute of Mathematical Sciences New York University - New York, NY 10003 Abstract
More informationAdvanced Introduction to Machine Learning, CMU-10715
Advanced Introduction to Machine Learning, CMU-10715 Deep Learning Barnabás Póczos, Sept 17 Credits Many of the pictures, results, and other materials are taken from: Ruslan Salakhutdinov Joshua Bengio
More informationMachine Learning With Python. Bin Chen Nov. 7, 2017 Research Computing Center
Machine Learning With Python Bin Chen Nov. 7, 2017 Research Computing Center Outline Introduction to Machine Learning (ML) Introduction to Neural Network (NN) Introduction to Deep Learning NN Introduction
More informationJOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS. Puyang Xu, Ruhi Sarikaya. Microsoft Corporation
JOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS Puyang Xu, Ruhi Sarikaya Microsoft Corporation ABSTRACT We describe a joint model for intent detection and slot filling based
More informationChannel Locality Block: A Variant of Squeeze-and-Excitation
Channel Locality Block: A Variant of Squeeze-and-Excitation 1 st Huayu Li Northern Arizona University Flagstaff, United State Northern Arizona University hl459@nau.edu arxiv:1901.01493v1 [cs.lg] 6 Jan
More informationDeep Fusion of Remote Sensing Data for Accurate Classification
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS 1 Deep Fusion of Remote Sensing Data for Accurate Classification Yushi Chen, Chunyang Li, Pedram Ghamisi, Xiuping Jia, Yanfeng Gu Abstract The multisensory fusion
More informationCPSC340. State-of-the-art Neural Networks. Nando de Freitas November, 2012 University of British Columbia
CPSC340 State-of-the-art Neural Networks Nando de Freitas November, 2012 University of British Columbia Outline of the lecture This lecture provides an overview of two state-of-the-art neural networks:
More informationHandwritten Hindi Numerals Recognition System
CS365 Project Report Handwritten Hindi Numerals Recognition System Submitted by: Akarshan Sarkar Kritika Singh Project Mentor: Prof. Amitabha Mukerjee 1 Abstract In this project, we consider the problem
More informationData Mining. Neural Networks
Data Mining Neural Networks Goals for this Unit Basic understanding of Neural Networks and how they work Ability to use Neural Networks to solve real problems Understand when neural networks may be most
More informationLSTM: An Image Classification Model Based on Fashion-MNIST Dataset
LSTM: An Image Classification Model Based on Fashion-MNIST Dataset Kexin Zhang, Research School of Computer Science, Australian National University Kexin Zhang, U6342657@anu.edu.au Abstract. The application
More informationWeighted Convolutional Neural Network. Ensemble.
Weighted Convolutional Neural Network Ensemble Xavier Frazão and Luís A. Alexandre Dept. of Informatics, Univ. Beira Interior and Instituto de Telecomunicações Covilhã, Portugal xavierfrazao@gmail.com
More informationFuzzy Entropy based feature selection for classification of hyperspectral data
Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering NIT Kurukshetra, 136119 mpce_pal@yahoo.co.uk Abstract: This paper proposes to use
More informationDEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla
DEEP LEARNING REVIEW Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature 2015 -Presented by Divya Chitimalla What is deep learning Deep learning allows computational models that are composed of multiple
More informationMachine Learning in WAN Research
Machine Learning in WAN Research Mariam Kiran mkiran@es.net Energy Sciences Network (ESnet) Lawrence Berkeley National Lab Oct 2017 Presented at Internet2 TechEx 2017 Outline ML in general ML in network
More informationSVD-based Universal DNN Modeling for Multiple Scenarios
SVD-based Universal DNN Modeling for Multiple Scenarios Changliang Liu 1, Jinyu Li 2, Yifan Gong 2 1 Microsoft Search echnology Center Asia, Beijing, China 2 Microsoft Corporation, One Microsoft Way, Redmond,
More informationDeep Learning. Volker Tresp Summer 2014
Deep Learning Volker Tresp Summer 2014 1 Neural Network Winter and Revival While Machine Learning was flourishing, there was a Neural Network winter (late 1990 s until late 2000 s) Around 2010 there
More informationAn Analysis of Single-Layer Networks in Unsupervised Feature Learning
An Analysis of Single-Layer Networks in Unsupervised Feature Learning Adam Coates Honglak Lee Andrew Y. Ng Stanford University Computer Science Dept. 353 Serra Mall Stanford, CA 94305 University of Michigan
More informationEnergy Based Models, Restricted Boltzmann Machines and Deep Networks. Jesse Eickholt
Energy Based Models, Restricted Boltzmann Machines and Deep Networks Jesse Eickholt ???? Who s heard of Energy Based Models (EBMs) Restricted Boltzmann Machines (RBMs) Deep Belief Networks Auto-encoders
More informationDeep Neural Networks for Recognizing Online Handwritten Mathematical Symbols
Deep Neural Networks for Recognizing Online Handwritten Mathematical Symbols Hai Dai Nguyen 1, Anh Duc Le 2 and Masaki Nakagawa 3 Tokyo University of Agriculture and Technology 2-24-16 Nakacho, Koganei-shi,
More informationReal-time Object Detection CS 229 Course Project
Real-time Object Detection CS 229 Course Project Zibo Gong 1, Tianchang He 1, and Ziyi Yang 1 1 Department of Electrical Engineering, Stanford University December 17, 2016 Abstract Objection detection
More informationLearning visual odometry with a convolutional network
Learning visual odometry with a convolutional network Kishore Konda 1, Roland Memisevic 2 1 Goethe University Frankfurt 2 University of Montreal konda.kishorereddy@gmail.com, roland.memisevic@gmail.com
More informationTHE MNIST DATABASE of handwritten digits Yann LeCun, Courant Institute, NYU Corinna Cortes, Google Labs, New York
THE MNIST DATABASE of handwritten digits Yann LeCun, Courant Institute, NYU Corinna Cortes, Google Labs, New York The MNIST database of handwritten digits, available from this page, has a training set
More informationDeep Learning in Visual Recognition. Thanks Da Zhang for the slides
Deep Learning in Visual Recognition Thanks Da Zhang for the slides Deep Learning is Everywhere 2 Roadmap Introduction Convolutional Neural Network Application Image Classification Object Detection Object
More informationStacked Denoising Autoencoders for Face Pose Normalization
Stacked Denoising Autoencoders for Face Pose Normalization Yoonseop Kang 1, Kang-Tae Lee 2,JihyunEun 2, Sung Eun Park 2 and Seungjin Choi 1 1 Department of Computer Science and Engineering Pohang University
More informationCOMP 551 Applied Machine Learning Lecture 16: Deep Learning
COMP 551 Applied Machine Learning Lecture 16: Deep Learning Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted, all
More informationArtificial Neural Networks. Introduction to Computational Neuroscience Ardi Tampuu
Artificial Neural Networks Introduction to Computational Neuroscience Ardi Tampuu 7.0.206 Artificial neural network NB! Inspired by biology, not based on biology! Applications Automatic speech recognition
More informationText Detection and Character Recognition in Scene Images with Unsupervised Feature Learning
Text Detection and Character Recognition in Scene Images with Unsupervised Feature Learning Adam Coates, Blake Carpenter, Carl Case, Sanjeev Satheesh, Bipin Suresh, Tao Wang, Andrew Y. Ng Computer Science
More informationDeep Convolutional Neural Networks. Nov. 20th, 2015 Bruce Draper
Deep Convolutional Neural Networks Nov. 20th, 2015 Bruce Draper Background: Fully-connected single layer neural networks Feed-forward classification Trained through back-propagation Example Computer Vision
More information3D model classification using convolutional neural network
3D model classification using convolutional neural network JunYoung Gwak Stanford jgwak@cs.stanford.edu Abstract Our goal is to classify 3D models directly using convolutional neural network. Most of existing
More informationRadial Basis Function Neural Network Classifier
Recognition of Unconstrained Handwritten Numerals by a Radial Basis Function Neural Network Classifier Hwang, Young-Sup and Bang, Sung-Yang Department of Computer Science & Engineering Pohang University
More informationLearning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li
Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,
More informationDeep Learning. Deep Learning provided breakthrough results in speech recognition and image classification. Why?
Data Mining Deep Learning Deep Learning provided breakthrough results in speech recognition and image classification. Why? Because Speech recognition and image classification are two basic examples of
More informationConvolutional Deep Belief Networks on CIFAR-10
Convolutional Deep Belief Networks on CIFAR-10 Alex Krizhevsky kriz@cs.toronto.edu 1 Introduction We describe how to train a two-layer convolutional Deep Belief Network (DBN) on the 1.6 million tiny images
More informationKernels vs. DNNs for Speech Recognition
Kernels vs. DNNs for Speech Recognition Joint work with: Columbia: Linxi (Jim) Fan, Michael Collins (my advisor) USC: Zhiyun Lu, Kuan Liu, Alireza Bagheri Garakani, Dong Guo, Aurélien Bellet, Fei Sha IBM:
More informationQuo Vadis, Action Recognition? A New Model and the Kinetics Dataset. By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018
Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018 Outline: Introduction Action classification architectures
More informationFacial Expression Classification with Random Filters Feature Extraction
Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle
More informationCOMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017
COMP9444 Neural Networks and Deep Learning 7. Image Processing COMP9444 17s2 Image Processing 1 Outline Image Datasets and Tasks Convolution in Detail AlexNet Weight Initialization Batch Normalization
More informationCS6220: DATA MINING TECHNIQUES
CS6220: DATA MINING TECHNIQUES Image Data: Classification via Neural Networks Instructor: Yizhou Sun yzsun@ccs.neu.edu November 19, 2015 Methods to Learn Classification Clustering Frequent Pattern Mining
More informationGPU-Accelerated Deep Learning
GPU-Accelerated Deep Learning July 6 th, 2016. Greg Heinrich. Credits: Alison B. Lowndes, Julie Bernauer, Leo K. Tam. PRACTICAL DEEP LEARNING EXAMPLES Image Classification, Object Detection, Localization,
More informationA Two-phase Distributed Training Algorithm for Linear SVM in WSN
Proceedings of the World Congress on Electrical Engineering and Computer Systems and Science (EECSS 015) Barcelona, Spain July 13-14, 015 Paper o. 30 A wo-phase Distributed raining Algorithm for Linear
More informationarxiv: v1 [cs.cv] 29 Oct 2017
A SAAK TRANSFORM APPROACH TO EFFICIENT, SCALABLE AND ROBUST HANDWRITTEN DIGITS RECOGNITION Yueru Chen, Zhuwei Xu, Shanshan Cai, Yujian Lang and C.-C. Jay Kuo Ming Hsieh Department of Electrical Engineering
More informationUsing Capsule Networks. for Image and Speech Recognition Problems. Yan Xiong
Using Capsule Networks for Image and Speech Recognition Problems by Yan Xiong A Thesis Presented in Partial Fulfillment of the Requirements for the Degree Master of Science Approved November 2018 by the
More informationConvolution Neural Networks for Chinese Handwriting Recognition
Convolution Neural Networks for Chinese Handwriting Recognition Xu Chen Stanford University 450 Serra Mall, Stanford, CA 94305 xchen91@stanford.edu Abstract Convolutional neural networks have been proven
More informationTiny ImageNet Visual Recognition Challenge
Tiny ImageNet Visual Recognition Challenge Ya Le Department of Statistics Stanford University yle@stanford.edu Xuan Yang Department of Electrical Engineering Stanford University xuany@stanford.edu Abstract
More informationCS 179 Lecture 16. Logistic Regression & Parallel SGD
CS 179 Lecture 16 Logistic Regression & Parallel SGD 1 Outline logistic regression (stochastic) gradient descent parallelizing SGD for neural nets (with emphasis on Google s distributed neural net implementation)
More information