Frequent Inner-Class Approach: A Semi-supervised Learning Technique for One-shot Learning
|
|
- Jonah Walsh
- 5 years ago
- Views:
Transcription
1 Frequent Inner-Class Approach: A Semi-supervised Learning Technique for One-shot Learning Izumi Suzuki, Koich Yamada, Muneyuki Unehara Nagaoka University of Technology, , Kamitomioka Nagaoka, Niigata , JAPAN {suzuki, yamada, unehara}@kjs.nagaokaut.ac.jp Abstract. The frequent inner-class (FIC) approach is to obtain knowledge from both supervised and unsupervised learning to recognize objects based on a small number of labeled training data, one of the implementation of one-shot learning. Irrespective of the features used to represent the data, it is necessary for FIC approach to (1) represent each data item not as a point of feature space but as various points of feature space, and (2) acquire knowledge from training data to learn useful representations of the data. In this research, each data representation is regarded as a class, or more specifically, an inner class. Also, in this paper, any training data are assumed to constitute a set of objects, each of them belonging to a class. Then, as the supervised learning proceeds, it modifies the usefulness value of each inner class and generates a higher-level inner class that consists of classes related to each other. This technique can be applied to many types of image features, including reduction of vector dimensions by neural networks. Also, recognition of classes by FIC approach is an extension of the formal representation of training data. Keywords: Classification, Data representation, Co-occurrence, Feature space, Image recognition by humans 1 Introduction Image recognition, as well as image classification, requires a large quantity of training data. Even more data are required when objects in the class have large variation. Meanwhile, humans sometimes recognize the class of objects even if the number of labeled data is very small or even one, especially when a person gets used to seeing the object. For example, a person who does not know tulip will recognize it as a flower by training with only one labeled data item, if the person gets used to seeing the flower shape of the tulip. This is known as one-shot learning [1], in computer vision, aims to learn information about object classes from a small number of training data, by transferring knowledge [2] of intra-class variabilities of learnt classes through model parameters, sharing parts, or contextual information. The intra-class variabilities include object variations by lightning changes, viewpoint changes, and occlusions. This paper introduces the frequent inner-class (FIC) approach for oneshot learning carrying out by collecting a large number of data representations (inner 859
2 classes) that occur frequently on numerous unlabeled data, and by identifying the best matched inner classes to the unknown object. The features useful for computer vision may not always be the same as those that a human would perceive from the object. However, this difference does not invalidate the following two requirements of human recognition on FIC approach. First, each object is represented not as a point of feature space but as various points of feature space. This is not the same as having many types of features or detecting many local features in one object. Each point of feature space is referred to as a data representation and corresponds to a class. For example, from Figure 1(a), a human can perceive various data representations, or classes, such as By using object segmentation, (1) calculator or (2) keychain; By outlines, (3) rectangle, (4) rounded rectangle, or (5) button (a shape with more detail); By texture, (6) metallic; and By combining data representations, (7) class of combination calculator and keychain or (8) class of calculator and keyring combined in a specific way. Second, simply collecting data representations from unlabeled images is not enough for FIC approach. Knowledge must first be acquired from various training data to select useful data representations. From the above discussion about Figure 1(a), classes 3 and 6 are too common and classes 5 and 8 are too rare, giving these classes little chance of being useful; meanwhile, the other classes have a chance to be useful data representations. Class 1, calculator, is selected if both images in Figure 1 are assigned to the same class. In this way, the FIC approach is a semi-supervised learning [2] that uses both unlabeled data and training data, such as labeled data. Also, the FIC approach is an application of feature learning [3], in the sense that it discovers useful features. (a) (b) Fig. 1. Object (a) has aspects of at least two classes: calculator and keychain. (All the images used in this paper are from the Caltech-256 data set [4].) 860
3 2 Frequent Inner-class (FIC) Approach One data point on a feature space is referred to as data representation. Given unlabeled data and training data, FIC approach is a semi-supervised learning to discover more useful data representations from various features of an object. The following are three examples of training data: 1. Objects of the same class. The class label is not necessary. This example includes various appearances of the same object caused by different views or illumination or a group of similar but different objects, as in Figure 2(a-c), as well as dissimilar objects of the same class, as in Figure 2(d). 2. Suggested data representations. For example, a labeled image of lenticular cloud suggests an important data representation for its class label. 3. Rare but important objects. For example, a person can regard an animal as dangerous and important owing to experience if the person was bitten by the animal. Only the first example is considered in this paper. 3 Related Techniques The purpose of proposing this technique is to optimize data representation, and it can be applied to image classifications such as classification performed by a support vector machine (SVM). As a matter of fact, the recognized classes of FIC approach can be regarded as an extension of formal data representation of training data; i.e., one object corresponds to one data representation. The difference between the extraction of many data representations in this technique and the extraction of many local features, such as scale-invariant feature transform (SIFT) [5], is that the local features are transformed to a data representation by a pooling process [6], such as making a histogram using the bag-of-features (BOF) method [7]. Convolutional neural networks (CNNs) [8] also explore optimal data representation from among various data representations of an object by optimizing network parameters. However, optimized parameters of CNN are parameters specialized for representing the objects of several classes. On the other hand, applying importance values to classes in FIC approach is equivalent to the use of weights in text data processing, of which the term frequency inverse document frequency (TF-IDF) method is a well-known example. In fact, the learning by unlabeled data determines the importance value, which corresponds to the term frequency, and the learning by labeled data determines other importance values, which correspond to the inverse document frequency. The semi-supervised learning and FIC approach are different in their purpose. The purpose of semi-supervised learning is to draw the class boundary onto the feature space, whereas the purpose of FIC approach is to add weight to data representations on the feature space. By combining data representations, the co-occurrence of features can be taken into consideration. Part-based models also consider the co-occurrence of features. For example, the constellation model [9] represents the object face by describing the 861
4 spatial assignment of parts, such as eye, mouth, and nose. Nowozin et al. [10] proposed a technique to find the best combination of visual words for image processing. A more recent example is the sparse visual bigram [11], which represents the co-occurrence of any pair of neighboring visual words. 4 Feature Space The feature space has no particular restriction except that 1) each object must be represented not as a point of feature space but as various points of feature space, and that 2) a data representation is made by combining or spatially assigning more than one data representation. If a data representation is made by more than one data representation, then data representations can be almost endlessly generated from one object. Although it ignores any spatial information, the BOF method satisfies this condition. Dimension reductions by neural networks, such as CNNs, which form data representations from pixelated images by repeated pooling and sampling processes, also satisfy this condition. This is the same as forming a data representation by simply scaling down the pixelated image. Designing the feature space more effectively for the FIC approach is a challenge for the future. 5 Method for FIC Approach Unsupervised learning and supervised learning can be carried out in parallel using the following overall approach. Knowledge The data representations are collected by unsupervised learning and then quantized by combining similar representations to make inner classes. The collection of inner classes with their usefulness values is the knowledge to be acquired. Inner classes are low level if they occur frequently on unlabeled data or high level if they occur rarely on unlabeled data. Unsupervised Learning (with unlabeled data) Extract all (or, depending on the case, sampled) data representations from each data item to create the knowledge. Higher-level inner classes are given a higher usefulness value. Extremely high-level inner classes are deleted, since they are less useful. 862
5 Supervised Learning (with training data, carried out with the recognition process) The usefulness value of the recognized inner classes is increased, while the usefulness values of other classes are slightly decreased. If more than one class is recognized, Generate a higher-level inner class equal to the set of recognized classes. In this generation, the related classes, such as co-occurring classes, are kept track of. If a label is pasted on the training data, then paste the same label on recognized inner classes. Delete extremely useless inner classes. Recognition When the data constitute one object, Extract data representations from the unknown object, and then list all the matched inner classes. Regard some of the most useful (and therefore, perhaps, higher-level) inner classes from among matched inner classes as the recognized classes. (The process that decides one class is not discussed in this paper; perhaps humans make the decision using other information, such as the surroundings.) When the data contain more than one object, If a class is common to every object, then this class is the recognized class and this set of objects is said to be recognizable. If the data are not recognizable, Make subsets of the objects so that each of the subsets is recognizable and that the union of the subsets equals the data. For example, make the subsets {a, b}, {b, c}, and {d} for the objects in Figure 2. Another condition might be required for making subsets; however, the authors are still investigating this. The recognized classes to be output are (1) the recognized classes for each subset and (2) the subsets, i.e., the inner classes. 6 Extended Training Data for Classifier During the recognition process, when the data contain more than one object, the case where two objects do not share a common subset can be interpreted as the impressions are different between two objects even though they belong to the same class. Examples of this are (1) the objects a and c and (2) the object d and the other objects in Figure 2. This case also arises if the feature space design is inadequate. However, the recognized classes are still useful as training data. This is because the formal representation of training data, where each object has only one data representa- 863
6 tion, is made by taking obvious singleton sets as the subsets, and by choosing one recognized class in each singleton set (see Figure 3). (a) (b) (c) (d) Fig. 2. Objects of the class mailbox. The impressions are different between (d) and the rest. Fig. 3. Illustration of the recognized classes in each object of Fig. 2. Recognized classes are marked by (a), (b), (c), and (d). Hatched symbols are the recognized classes of {a, b} and {b, c}. The formal representation of training data, where one data representation corresponds to one object, is marked by a checkmark ( ). If the feature space design is inadequate, then some subsets, {b, c} for example, may not be recognizable. Instead, the data representation turns back to the formal representations, perhaps the checkmarks. 864
7 7 Simulation We are preparing a simulation to check the effectiveness of FIC approach. The results will be reported as soon as possible. Actual images are not used in the model. Instead, data representations are synthesized, and then various data representations are generated by combining the representations. As a matter of course, spatial information is not considered in the model: The feature space is composed of (1) N elements, referred to as visual-words, as elementary classes and (2) less-than-or-equal-to M-combinations from the set of N elements. Therefore, the cardinality of feature space is MM mm=1 CC(NN, mm), where C(N,m) is the number of m-combinations from the set of N elements. An image comprises R different visual-words and their combinations. Whether two images belong to the same class is determined by the number of visual-words common to the two images. The following issues are to be examined by the simulation. How the deletion of less useful inner classes affects the performance. Deletion is necessary to prevent overexpansion of the number of inner classes. The effectiveness of supervised learning, that is, whether the recognition rate increases for objects belonging to the class already learned using training data. Furthermore, the authors expect that the recognition rate also increases for the objects of classes that are not learned yet. The best order of learning by comparing two modes: (1) unsupervised learning is separately followed by supervised learning and (2) unsupervised and supervised learning are carried out in parallel. What kind of training data should be learned, and in what order should it be learned? For example, does learning the same data repeatedly affect performance? 8 Conclusion This paper introduced frequent inner-class (FIC) approach for one-shot learning, the recognition of an object with few labeled data by acquiring knowledge, in advance, from a large amount of unlabeled data. This technique is one of the implementation of one-shot learning. Because knowledge must be acquired from training data to select useful features, the FIC approach is semi-supervised learning. This technique is expected to improve the performance of image classifiers, such as SVM, since the recognized classes of FIC approach can be regarded as an extension of formal data representation, i.e., one object corresponding to one data representation. The authors intend to carry out the simulation as soon as possible and then perform experiments with actual images. Our priority is to design the feature space effectively, which is key to the FIC approach of one-shot learning. 865
8 References 1. Fei-Fei, L., Fergus, R., Perona, P.: One-Shot learning of object categories. IEEE Trans. Pattern Analysis and Machine Intelligence, Vol. 28(4), pp , (2006) 2. Miller, E., Matsakis, N., Viola, P.: Learning from one example through shared densities on transforms. In CVPR, v(1), pp , (2000) 3. Chapelle, O., Schölkopf, B., Zien, A.: Semi-Supervised Learning. MIT Press (2006) 4. Bengio, Y., Courville, A., Vincent, P.: Representation Learning: A Review and New Perspectives. IEEE Trans. PAMI, special issue Learning Deep Architectures. 35, pp (2013) 5. The Caltech 256. Web. May 2017, 6. Lowe, D. G.: Distinctive Image Features from Scale-invariant Keypoints. Int. Journal of Computer Vision, Vol. 60, No. 2, pp , (2004) 7. Boureau, Y.L., Bach, F., LeCun, Y., Ponce, J.: Learning Mid-Level Features For Recognition. In: CVPR (2010) 8. Csurka, G., Bray, C., Dance, C., Fan, L.: Visual Categorization with Bags of Keypoints. In: Workshop on Statistical Learning in Computer Vision, European Conference on Computer Vision, pp (2004) 9. LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradientbased learning applied to document recognition. In: Proceedings of the IEEE, 86(11), pp (1998) 10. Weber, M., Welling, M., Perona, P.: Unsupervised Learning of Models for Recognition. In: Proc. of the 6th European Conference on Computer Vision (ECCV 00), Part I, pp (2000) 11. Nowozin, S., Tsuda, K., Uno, T., Kudo, T., Bakir, G.: Weighted Substructure Mining for Image Analysis. In: CVPR 07 (2007) 12. Jiang, Y. G., Yang, J., Ngo, C.W., Hauptmann, A.G.: Representations of Keypoint-Based Semantic Concept Detection: A Comprehensive Study. IEEE Trans. On Multimedia 12(1), pp (2010) 866
String distance for automatic image classification
String distance for automatic image classification Nguyen Hong Thinh*, Le Vu Ha*, Barat Cecile** and Ducottet Christophe** *University of Engineering and Technology, Vietnam National University of HaNoi,
More informationImproving Recognition through Object Sub-categorization
Improving Recognition through Object Sub-categorization Al Mansur and Yoshinori Kuno Graduate School of Science and Engineering, Saitama University, 255 Shimo-Okubo, Sakura-ku, Saitama-shi, Saitama 338-8570,
More informationBeyond Bags of Features
: for Recognizing Natural Scene Categories Matching and Modeling Seminar Instructed by Prof. Haim J. Wolfson School of Computer Science Tel Aviv University December 9 th, 2015
More informationBag of Words Models. CS4670 / 5670: Computer Vision Noah Snavely. Bag-of-words models 11/26/2013
CS4670 / 5670: Computer Vision Noah Snavely Bag-of-words models Object Bag of words Bag of Words Models Adapted from slides by Rob Fergus and Svetlana Lazebnik 1 Object Bag of words Origin 1: Texture Recognition
More informationBeyond bags of features: Adding spatial information. Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba
Beyond bags of features: Adding spatial information Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba Adding spatial information Forming vocabularies from pairs of nearby features doublets
More informationPreviously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011
Previously Part-based and local feature models for generic object recognition Wed, April 20 UT-Austin Discriminative classifiers Boosting Nearest neighbors Support vector machines Useful for object recognition
More informationRushes Video Segmentation Using Semantic Features
Rushes Video Segmentation Using Semantic Features Athina Pappa, Vasileios Chasanis, and Antonis Ioannidis Department of Computer Science and Engineering, University of Ioannina, GR 45110, Ioannina, Greece
More informationImgSeek: Capturing User s Intent For Internet Image Search
ImgSeek: Capturing User s Intent For Internet Image Search Abstract - Internet image search engines (e.g. Bing Image Search) frequently lean on adjacent text features. It is difficult for them to illustrate
More informationarxiv: v1 [cs.lg] 20 Dec 2013
Unsupervised Feature Learning by Deep Sparse Coding Yunlong He Koray Kavukcuoglu Yun Wang Arthur Szlam Yanjun Qi arxiv:1312.5783v1 [cs.lg] 20 Dec 2013 Abstract In this paper, we propose a new unsupervised
More informationExtracting Spatio-temporal Local Features Considering Consecutiveness of Motions
Extracting Spatio-temporal Local Features Considering Consecutiveness of Motions Akitsugu Noguchi and Keiji Yanai Department of Computer Science, The University of Electro-Communications, 1-5-1 Chofugaoka,
More informationLecture 16: Object recognition: Part-based generative models
Lecture 16: Object recognition: Part-based generative models Professor Stanford Vision Lab 1 What we will learn today? Introduction Constellation model Weakly supervised training One-shot learning (Problem
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 16: Bag-of-words models Object Bag of words Announcements Project 3: Eigenfaces due Wednesday, November 11 at 11:59pm solo project Final project presentations:
More informationAggregating Descriptors with Local Gaussian Metrics
Aggregating Descriptors with Local Gaussian Metrics Hideki Nakayama Grad. School of Information Science and Technology The University of Tokyo Tokyo, JAPAN nakayama@ci.i.u-tokyo.ac.jp Abstract Recently,
More informationPart based models for recognition. Kristen Grauman
Part based models for recognition Kristen Grauman UT Austin Limitations of window-based models Not all objects are box-shaped Assuming specific 2d view of object Local components themselves do not necessarily
More informationEfficient Kernels for Identifying Unbounded-Order Spatial Features
Efficient Kernels for Identifying Unbounded-Order Spatial Features Yimeng Zhang Carnegie Mellon University yimengz@andrew.cmu.edu Tsuhan Chen Cornell University tsuhan@ece.cornell.edu Abstract Higher order
More informationPart-based and local feature models for generic object recognition
Part-based and local feature models for generic object recognition May 28 th, 2015 Yong Jae Lee UC Davis Announcements PS2 grades up on SmartSite PS2 stats: Mean: 80.15 Standard Dev: 22.77 Vote on piazza
More informationDiscriminative Object Class Models of Appearance and Shape by Correlatons
Discriminative Object Class Models of Appearance and Shape by Correlatons S. Savarese, J. Winn, A. Criminisi University of Illinois at Urbana-Champaign Microsoft Research Ltd., Cambridge, CB3 0FB, United
More informationSupervised learning. y = f(x) function
Supervised learning y = f(x) output prediction function Image feature Training: given a training set of labeled examples {(x 1,y 1 ),, (x N,y N )}, estimate the prediction function f by minimizing the
More informationDiscovering Visual Hierarchy through Unsupervised Learning Haider Razvi
Discovering Visual Hierarchy through Unsupervised Learning Haider Razvi hrazvi@stanford.edu 1 Introduction: We present a method for discovering visual hierarchy in a set of images. Automatically grouping
More informationObject Recognition. Computer Vision. Slides from Lana Lazebnik, Fei-Fei Li, Rob Fergus, Antonio Torralba, and Jean Ponce
Object Recognition Computer Vision Slides from Lana Lazebnik, Fei-Fei Li, Rob Fergus, Antonio Torralba, and Jean Ponce How many visual object categories are there? Biederman 1987 ANIMALS PLANTS OBJECTS
More informationLearning Visual Semantics: Models, Massive Computation, and Innovative Applications
Learning Visual Semantics: Models, Massive Computation, and Innovative Applications Part II: Visual Features and Representations Liangliang Cao, IBM Watson Research Center Evolvement of Visual Features
More informationLatest development in image feature representation and extraction
International Journal of Advanced Research and Development ISSN: 2455-4030, Impact Factor: RJIF 5.24 www.advancedjournal.com Volume 2; Issue 1; January 2017; Page No. 05-09 Latest development in image
More informationDeep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks
Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Si Chen The George Washington University sichen@gwmail.gwu.edu Meera Hahn Emory University mhahn7@emory.edu Mentor: Afshin
More informationSelection of Scale-Invariant Parts for Object Class Recognition
Selection of Scale-Invariant Parts for Object Class Recognition Gy. Dorkó and C. Schmid INRIA Rhône-Alpes, GRAVIR-CNRS 655, av. de l Europe, 3833 Montbonnot, France fdorko,schmidg@inrialpes.fr Abstract
More informationTRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK
TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK 1 Po-Jen Lai ( 賴柏任 ), 2 Chiou-Shann Fuh ( 傅楸善 ) 1 Dept. of Electrical Engineering, National Taiwan University, Taiwan 2 Dept.
More informationGeneric object recognition using graph embedding into a vector space
American Journal of Software Engineering and Applications 2013 ; 2(1) : 13-18 Published online February 20, 2013 (http://www.sciencepublishinggroup.com/j/ajsea) doi: 10.11648/j. ajsea.20130201.13 Generic
More informationCodebook Graph Coding of Descriptors
Int'l Conf. Par. and Dist. Proc. Tech. and Appl. PDPTA'5 3 Codebook Graph Coding of Descriptors Tetsuya Yoshida and Yuu Yamada Graduate School of Humanities and Science, Nara Women s University, Nara,
More informationCost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling
[DOI: 10.2197/ipsjtcva.7.99] Express Paper Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling Takayoshi Yamashita 1,a) Takaya Nakamura 1 Hiroshi Fukui 1,b) Yuji
More informationArtistic ideation based on computer vision methods
Journal of Theoretical and Applied Computer Science Vol. 6, No. 2, 2012, pp. 72 78 ISSN 2299-2634 http://www.jtacs.org Artistic ideation based on computer vision methods Ferran Reverter, Pilar Rosado,
More informationAnnouncements. Recognition. Recognition. Recognition. Recognition. Homework 3 is due May 18, 11:59 PM Reading: Computer Vision I CSE 152 Lecture 14
Announcements Computer Vision I CSE 152 Lecture 14 Homework 3 is due May 18, 11:59 PM Reading: Chapter 15: Learning to Classify Chapter 16: Classifying Images Chapter 17: Detecting Objects in Images Given
More informationObject Category Detection. Slides mostly from Derek Hoiem
Object Category Detection Slides mostly from Derek Hoiem Today s class: Object Category Detection Overview of object category detection Statistical template matching with sliding window Part-based Models
More informationA Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images
A Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images Marc Aurelio Ranzato Yann LeCun Courant Institute of Mathematical Sciences New York University - New York, NY 10003 Abstract
More informationVideo annotation based on adaptive annular spatial partition scheme
Video annotation based on adaptive annular spatial partition scheme Guiguang Ding a), Lu Zhang, and Xiaoxu Li Key Laboratory for Information System Security, Ministry of Education, Tsinghua National Laboratory
More informationSemantic-based image analysis with the goal of assisting artistic creation
Semantic-based image analysis with the goal of assisting artistic creation Pilar Rosado 1, Ferran Reverter 2, Eva Figueras 1, and Miquel Planas 1 1 Fine Arts Faculty, University of Barcelona, Spain, pilarrosado@ub.edu,
More informationA Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images
A Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images Marc Aurelio Ranzato Yann LeCun Courant Institute of Mathematical Sciences New York University - New York, NY 10003 Abstract
More informationTransfer Forest Based on Covariate Shift
Transfer Forest Based on Covariate Shift Masamitsu Tsuchiya SECURE, INC. tsuchiya@secureinc.co.jp Yuji Yamauchi, Takayoshi Yamashita, Hironobu Fujiyoshi Chubu University yuu@vision.cs.chubu.ac.jp, {yamashita,
More informationPart Localization by Exploiting Deep Convolutional Networks
Part Localization by Exploiting Deep Convolutional Networks Marcel Simon, Erik Rodner, and Joachim Denzler Computer Vision Group, Friedrich Schiller University of Jena, Germany www.inf-cv.uni-jena.de Abstract.
More informationUnsupervised Identification of Multiple Objects of Interest from Multiple Images: discover
Unsupervised Identification of Multiple Objects of Interest from Multiple Images: discover Devi Parikh and Tsuhan Chen Carnegie Mellon University {dparikh,tsuhan}@cmu.edu Abstract. Given a collection of
More informationPreliminary Local Feature Selection by Support Vector Machine for Bag of Features
Preliminary Local Feature Selection by Support Vector Machine for Bag of Features Tetsu Matsukawa Koji Suzuki Takio Kurita :University of Tsukuba :National Institute of Advanced Industrial Science and
More informationSupervised learning. y = f(x) function
Supervised learning y = f(x) output prediction function Image feature Training: given a training set of labeled examples {(x 1,y 1 ),, (x N,y N )}, estimate the prediction function f by minimizing the
More informationVisual Object Recognition
Perceptual and Sensory Augmented Computing Visual Object Recognition Tutorial Visual Object Recognition Bastian Leibe Computer Vision Laboratory ETH Zurich Chicago, 14.07.2008 & Kristen Grauman Department
More informationImage Mining: frameworks and techniques
Image Mining: frameworks and techniques Madhumathi.k 1, Dr.Antony Selvadoss Thanamani 2 M.Phil, Department of computer science, NGM College, Pollachi, Coimbatore, India 1 HOD Department of Computer Science,
More informationCS4495/6495 Introduction to Computer Vision. 8C-L1 Classification: Discriminative models
CS4495/6495 Introduction to Computer Vision 8C-L1 Classification: Discriminative models Remember: Supervised classification Given a collection of labeled examples, come up with a function that will predict
More informationToward Part-based Document Image Decoding
2012 10th IAPR International Workshop on Document Analysis Systems Toward Part-based Document Image Decoding Wang Song, Seiichi Uchida Kyushu University, Fukuoka, Japan wangsong@human.ait.kyushu-u.ac.jp,
More informationFacial Expression Classification with Random Filters Feature Extraction
Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle
More informationDeep Neural Networks:
Deep Neural Networks: Part II Convolutional Neural Network (CNN) Yuan-Kai Wang, 2016 Web site of this course: http://pattern-recognition.weebly.com source: CNN for ImageClassification, by S. Lazebnik,
More informationA Survey on Image Classification using Data Mining Techniques Vyoma Patel 1 G. J. Sahani 2
IJSRD - International Journal for Scientific Research & Development Vol. 2, Issue 10, 2014 ISSN (online): 2321-0613 A Survey on Image Classification using Data Mining Techniques Vyoma Patel 1 G. J. Sahani
More informationA Novel Image Semantic Understanding and Feature Extraction Algorithm. and Wenzhun Huang
A Novel Image Semantic Understanding and Feature Extraction Algorithm Xinxin Xie 1, a 1, b* and Wenzhun Huang 1 School of Information Engineering, Xijing University, Xi an 710123, China a 346148500@qq.com,
More informationBeyond Bags of features Spatial information & Shape models
Beyond Bags of features Spatial information & Shape models Jana Kosecka Many slides adapted from S. Lazebnik, FeiFei Li, Rob Fergus, and Antonio Torralba Detection, recognition (so far )! Bags of features
More informationTensor Decomposition of Dense SIFT Descriptors in Object Recognition
Tensor Decomposition of Dense SIFT Descriptors in Object Recognition Tan Vo 1 and Dat Tran 1 and Wanli Ma 1 1- Faculty of Education, Science, Technology and Mathematics University of Canberra, Australia
More informationRobust Face Detection Based on Convolutional Neural Networks
Robust Face Detection Based on Convolutional Neural Networks M. Delakis and C. Garcia Department of Computer Science, University of Crete P.O. Box 2208, 71409 Heraklion, Greece {delakis, cgarcia}@csd.uoc.gr
More informationThe most cited papers in Computer Vision
COMPUTER VISION, PUBLICATION The most cited papers in Computer Vision In Computer Vision, Paper Talk on February 10, 2012 at 11:10 pm by gooly (Li Yang Ku) Although it s not always the case that a paper
More informationCombining Selective Search Segmentation and Random Forest for Image Classification
Combining Selective Search Segmentation and Random Forest for Image Classification Gediminas Bertasius November 24, 2013 1 Problem Statement Random Forest algorithm have been successfully used in many
More informationObject Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal
Object Detection Lecture 10.3 - Introduction to deep learning (CNN) Idar Dyrdal Deep Learning Labels Computational models composed of multiple processing layers (non-linear transformations) Used to learn
More informationAUTOMATIC VISUAL CONCEPT DETECTION IN VIDEOS
AUTOMATIC VISUAL CONCEPT DETECTION IN VIDEOS Nilam B. Lonkar 1, Dinesh B. Hanchate 2 Student of Computer Engineering, Pune University VPKBIET, Baramati, India Computer Engineering, Pune University VPKBIET,
More informationAn Efficient Approach for Color Pattern Matching Using Image Mining
An Efficient Approach for Color Pattern Matching Using Image Mining * Manjot Kaur Navjot Kaur Master of Technology in Computer Science & Engineering, Sri Guru Granth Sahib World University, Fatehgarh Sahib,
More informationLinear combinations of simple classifiers for the PASCAL challenge
Linear combinations of simple classifiers for the PASCAL challenge Nik A. Melchior and David Lee 16 721 Advanced Perception The Robotics Institute Carnegie Mellon University Email: melchior@cmu.edu, dlee1@andrew.cmu.edu
More informationComputer Vision Lecture 16
Computer Vision Lecture 16 Deep Learning for Object Categorization 14.01.2016 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar registration period
More informationComputer Vision & Digital Image Processing. Image segmentation: thresholding
Computer Vision & Digital Image Processing Image Segmentation: Thresholding Dr. D. J. Jackson Lecture 18-1 Image segmentation: thresholding Suppose an image f(y) is composed of several light objects on
More informationConvolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech
Convolutional Neural Networks Computer Vision Jia-Bin Huang, Virginia Tech Today s class Overview Convolutional Neural Network (CNN) Training CNN Understanding and Visualizing CNN Image Categorization:
More informationContent-Based Image Classification: A Non-Parametric Approach
1 Content-Based Image Classification: A Non-Parametric Approach Paulo M. Ferreira, Mário A.T. Figueiredo, Pedro M. Q. Aguiar Abstract The rise of the amount imagery on the Internet, as well as in multimedia
More informationGriMa: a Grid Mining Algorithm for Bag-of-Grid-Based Classification
GriMa: a Grid Mining Algorithm for Bag-of-Grid-Based Classification Romain Deville 1,2, Elisa Fromont 1, Baptiste Jeudy 1, and Christine Solnon 2 1 UJM, CNRS, LaHC UMR 5516, F-42000, SAINT-ETIENNE, France
More informationReal-time Object Detection CS 229 Course Project
Real-time Object Detection CS 229 Course Project Zibo Gong 1, Tianchang He 1, and Ziyi Yang 1 1 Department of Electrical Engineering, Stanford University December 17, 2016 Abstract Objection detection
More informationOBJECT CATEGORIZATION
OBJECT CATEGORIZATION Ing. Lorenzo Seidenari e-mail: seidenari@dsi.unifi.it Slides: Ing. Lamberto Ballan November 18th, 2009 What is an Object? Merriam-Webster Definition: Something material that may be
More informationImage Classification based on Saliency Driven Nonlinear Diffusion and Multi-scale Information Fusion Ms. Swapna R. Kharche 1, Prof.B.K.
Image Classification based on Saliency Driven Nonlinear Diffusion and Multi-scale Information Fusion Ms. Swapna R. Kharche 1, Prof.B.K.Chaudhari 2 1M.E. student, Department of Computer Engg, VBKCOE, Malkapur
More informationHW2 due on Thursday. Face Recognition: Dimensionality Reduction. Biometrics CSE 190 Lecture 11. Perceptron Revisited: Linear Separators
HW due on Thursday Face Recognition: Dimensionality Reduction Biometrics CSE 190 Lecture 11 CSE190, Winter 010 CSE190, Winter 010 Perceptron Revisited: Linear Separators Binary classification can be viewed
More informationFuzzy based Multiple Dictionary Bag of Words for Image Classification
Available online at www.sciencedirect.com Procedia Engineering 38 (2012 ) 2196 2206 International Conference on Modeling Optimisation and Computing Fuzzy based Multiple Dictionary Bag of Words for Image
More informationCategorization by Learning and Combining Object Parts
Categorization by Learning and Combining Object Parts Bernd Heisele yz Thomas Serre y Massimiliano Pontil x Thomas Vetter Λ Tomaso Poggio y y Center for Biological and Computational Learning, M.I.T., Cambridge,
More informationRecognition with Bag-ofWords. (Borrowing heavily from Tutorial Slides by Li Fei-fei)
Recognition with Bag-ofWords (Borrowing heavily from Tutorial Slides by Li Fei-fei) Recognition So far, we ve worked on recognizing edges Now, we ll work on recognizing objects We will use a bag-of-words
More informationDetecting Printed and Handwritten Partial Copies of Line Drawings Embedded in Complex Backgrounds
9 1th International Conference on Document Analysis and Recognition Detecting Printed and Handwritten Partial Copies of Line Drawings Embedded in Complex Backgrounds Weihan Sun, Koichi Kise Graduate School
More informationCELLULAR AUTOMATA BAG OF VISUAL WORDS FOR OBJECT RECOGNITION
U.P.B. Sci. Bull., Series C, Vol. 77, Iss. 4, 2015 ISSN 2286-3540 CELLULAR AUTOMATA BAG OF VISUAL WORDS FOR OBJECT RECOGNITION Ionuţ Mironică 1, Bogdan Ionescu 2, Radu Dogaru 3 In this paper we propose
More informationFace Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN
2016 International Conference on Artificial Intelligence: Techniques and Applications (AITA 2016) ISBN: 978-1-60595-389-2 Face Recognition Using Vector Quantization Histogram and Support Vector Machine
More informationCPPP/UFMS at ImageCLEF 2014: Robot Vision Task
CPPP/UFMS at ImageCLEF 2014: Robot Vision Task Rodrigo de Carvalho Gomes, Lucas Correia Ribas, Amaury Antônio de Castro Junior, Wesley Nunes Gonçalves Federal University of Mato Grosso do Sul - Ponta Porã
More informationDeformable Part Models
CS 1674: Intro to Computer Vision Deformable Part Models Prof. Adriana Kovashka University of Pittsburgh November 9, 2016 Today: Object category detection Window-based approaches: Last time: Viola-Jones
More informationAnnouncements. Recognition I. Gradient Space (p,q) What is the reflectance map?
Announcements I HW 3 due 12 noon, tomorrow. HW 4 to be posted soon recognition Lecture plan recognition for next two lectures, then video and motion. Introduction to Computer Vision CSE 152 Lecture 17
More informationFacial Expression Recognition Based on Local Directional Pattern Using SVM Decision-level Fusion
Facial Expression Recognition Based on Local Directional Pattern Using SVM Decision-level Fusion Juxiang Zhou 1, Tianwei Xu 2, Jianhou Gan 1 1. Key Laboratory of Education Informalization for Nationalities,
More informationPart-based models. Lecture 10
Part-based models Lecture 10 Overview Representation Location Appearance Generative interpretation Learning Distance transforms Other approaches using parts Felzenszwalb, Girshick, McAllester, Ramanan
More informationPattern recognition (3)
Pattern recognition (3) 1 Things we have discussed until now Statistical pattern recognition Building simple classifiers Supervised classification Minimum distance classifier Bayesian classifier Building
More informationCourse Administration
Course Administration Project 2 results are online Project 3 is out today The first quiz is a week from today (don t panic!) Covers all material up to the quiz Emphasizes lecture material NOT project topics
More informationINTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY
INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK REVIEW ON CONTENT BASED IMAGE RETRIEVAL BY USING VISUAL SEARCH RANKING MS. PRAGATI
More informationDEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla
DEEP LEARNING REVIEW Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature 2015 -Presented by Divya Chitimalla What is deep learning Deep learning allows computational models that are composed of multiple
More informationCENTERED FEATURES FROM INTEGRAL IMAGES
CENTERED FEATURES FROM INTEGRAL IMAGES ZOHAIB KHAN Bachelor of Applied Information Technology Thesis Report No. 2009:070 ISSN: 1651-4769 University of Gothenburg Department of Applied Information Technology
More informationAn evaluation of Nearest Neighbor Images to Classes versus Nearest Neighbor Images to Images Instructed by Professor David Jacobs Phil Huynh
An evaluation of Nearest Neighbor Images to Classes versus Nearest Neighbor Images to Images Instructed by Professor David Jacobs Phil Huynh Abstract In 2008, Boiman et al. announced in their work "In
More informationLEARNING TO GENERATE CHAIRS WITH CONVOLUTIONAL NEURAL NETWORKS
LEARNING TO GENERATE CHAIRS WITH CONVOLUTIONAL NEURAL NETWORKS Alexey Dosovitskiy, Jost Tobias Springenberg and Thomas Brox University of Freiburg Presented by: Shreyansh Daftry Visual Learning and Recognition
More informationMulti-Class Image Classification: Sparsity Does It Better
Multi-Class Image Classification: Sparsity Does It Better Sean Ryan Fanello 1,2, Nicoletta Noceti 2, Giorgio Metta 1 and Francesca Odone 2 1 Department of Robotics, Brain and Cognitive Sciences, Istituto
More informationEnsemble Image Classification Method Based on Genetic Image Network
Ensemble Image Classification Method Based on Genetic Image Network Shiro Nakayama, Shinichi Shirakawa, Noriko Yata and Tomoharu Nagao Graduate School of Environment and Information Sciences, Yokohama
More informationarxiv: v1 [cs.mm] 12 Jan 2016
Learning Subclass Representations for Visually-varied Image Classification Xinchao Li, Peng Xu, Yue Shi, Martha Larson, Alan Hanjalic Multimedia Information Retrieval Lab, Delft University of Technology
More informationFACE RECOGNITION FROM A SINGLE SAMPLE USING RLOG FILTER AND MANIFOLD ANALYSIS
FACE RECOGNITION FROM A SINGLE SAMPLE USING RLOG FILTER AND MANIFOLD ANALYSIS Jaya Susan Edith. S 1 and A.Usha Ruby 2 1 Department of Computer Science and Engineering,CSI College of Engineering, 2 Research
More informationBasic Problem Addressed. The Approach I: Training. Main Idea. The Approach II: Testing. Why a set of vocabularies?
Visual Categorization With Bags of Keypoints. ECCV,. G. Csurka, C. Bray, C. Dance, and L. Fan. Shilpa Gulati //7 Basic Problem Addressed Find a method for Generic Visual Categorization Visual Categorization:
More informationPart-Based Statistical Models for Object Classification and Detection
Part-Based Statistical Models for Object Classification and Detection Elliot Joel Bernstein and Yali Amit Department of Statistics, University of Chicago E-mail: {bernstei,amit}@galton.uchicago.edu Abstract
More informationCharacter Recognition Using Convolutional Neural Networks
Character Recognition Using Convolutional Neural Networks David Bouchain Seminar Statistical Learning Theory University of Ulm, Germany Institute for Neural Information Processing Winter 2006/2007 Abstract
More informationRecap Image Classification with Bags of Local Features
Recap Image Classification with Bags of Local Features Bag of Feature models were the state of the art for image classification for a decade BoF may still be the state of the art for instance retrieval
More informationFace Detection Using Convolutional Neural Networks and Gabor Filters
Face Detection Using Convolutional Neural Networks and Gabor Filters Bogdan Kwolek Rzeszów University of Technology W. Pola 2, 35-959 Rzeszów, Poland bkwolek@prz.rzeszow.pl Abstract. This paper proposes
More informationContent-Based Image Retrieval Using Deep Belief Networks
Content-Based Image Retrieval Using Deep Belief Networks By Jason Kroge Submitted to the graduate degree program in the Department of Electrical Engineering and Computer Science of the University of Kansas
More informationSelection of Scale-Invariant Parts for Object Class Recognition
Selection of Scale-Invariant Parts for Object Class Recognition Gyuri Dorkó, Cordelia Schmid To cite this version: Gyuri Dorkó, Cordelia Schmid. Selection of Scale-Invariant Parts for Object Class Recognition.
More informationPatch Descriptors. CSE 455 Linda Shapiro
Patch Descriptors CSE 455 Linda Shapiro How can we find corresponding points? How can we find correspondences? How do we describe an image patch? How do we describe an image patch? Patches with similar
More informationMACHINE VISION is a subfield of artificial intelligence. An Ensemble of Deep Support Vector Machines for Image Categorization
An Ensemble of Deep Support Vector Machines for Image Categorization Azizi Abdullah, Remco C. Veltkamp Department of Information and Computer Sciences Utrecht University, The Netherlands azizi@cs.uu.nl,
More informationIntegrated Feature Selection and Higher-order Spatial Feature Extraction for Object Categorization
Integrated Feature Selection and Higher-order Spatial Feature Extraction for Object Categorization David Liu, Gang Hua 2, Paul Viola 2, Tsuhan Chen Dept. of ECE, Carnegie Mellon University and Microsoft
More informationBayes Risk. Classifiers for Recognition Reading: Chapter 22 (skip 22.3) Discriminative vs Generative Models. Loss functions in classifiers
Classifiers for Recognition Reading: Chapter 22 (skip 22.3) Examine each window of an image Classify object class within each window based on a training set images Example: A Classification Problem Categorize
More information