A proposal of hierarchization method based on data distribution for multi class classification

Similar documents
Predicting Diabetes and Heart Disease Using Diagnostic Measurements and Supervised Learning Classification Models

Heart Disease Detection using EKSTRAP Clustering with Statistical and Distance based Classifiers

Performance Evaluation of Various Classification Algorithms

Package ordinalforest

Performance Comparison of Decision Tree Algorithms for Medical Data Sets

The Pseudo Gradient Search and a Penalty Technique Used in Classifications.

DATA MINING I - 1DL360

Report on Forensic Science Application

Artificial Neural Networks (Feedforward Nets)

Lecture 20: Bagging, Random Forests, Boosting

A neural network that classifies glass either as window or non-window depending on the glass chemistry.

Lecture 19: Decision trees

The Basics of Decision Trees

IJMIE Volume 2, Issue 9 ISSN:

International Journal of Scientific & Engineering Research, Volume 4, Issue 4, April ISSN

Class-Specific Feature Selection for One-Against-All Multiclass SVMs

Evolutionary Instance Selection Algorithm based on Takagi-Sugeno Fuzzy Model

A Dynamic Linkage Clustering using KD-Tree

Classification of Hand-Written Numeric Digits

Heart Disease Prediction and Classification Using Machine Learning Algorithms Optimized by Particle Swarm Optimization and Ant Colony Optimization

Machine Learning with MATLAB --classification

Predicting the Heart Attack Symptoms using Biomedical Data Mining Techniques

Rank Measures for Ordering

Learning Dot Product Polynomials for multiclass problems

Random Forests and Boosting

Why Do Nearest-Neighbour Algorithms Do So Well?

A genetic algorithm for interpretable model extraction from decision tree ensembles

Statistical Methods for Data Mining

Package FFTrees. November 2, 2017

Individual feature selection in each One-versus-One classifier improves multi-class SVM performance

FEATURE EXTRACTION TECHNIQUES USING SUPPORT VECTOR MACHINES IN DISEASE PREDICTION

Chapter 8 The C 4.5*stat algorithm

Learning highly non-separable Boolean functions using Constructive Feedforward Neural Network

Machine Learning - Clustering. CS102 Fall 2017

Fuzzy Signature Based Radial Basis Neural Network

Using Network Analysis to Improve Nearest Neighbor Classification of Non-Network Data

Mining di Dati Web. Lezione 3 - Clustering and Classification

Online Mathematical Symbol Recognition using SVMs with Features from Functional Approximation

Comparison of various classification models for making financial decisions

Classification/Regression Trees and Random Forests

Distance Weighted Discrimination Method for Parkinson s for Automatic Classification of Rehabilitative Speech Treatment for Parkinson s Patients

Feature weighting using particle swarm optimization for learning vector quantization classifier

A Systematic Overview of Data Mining Algorithms. Sargur Srihari University at Buffalo The State University of New York

The Effects of Outliers on Support Vector Machines

IMPLEMENTATION OF OWNERSHIP RIGHTS PROTECTION FOR NUMERIC AND NON-NUMERIC RELATIONAL DATA USING WATERMARKING

Using Decision Trees and Soft Labeling to Filter Mislabeled Data. Abstract

An Effective Performance of Feature Selection with Classification of Data Mining Using SVM Algorithm

Using a genetic algorithm for editing k-nearest neighbor classifiers

Performance Comparison of the Automatic Data Reduction System (ADRS)

Classification using Weka (Brain, Computation, and Neural Learning)

TWRBF Transductive RBF Neural Network with Weighted Data Normalization

Some questions of consensus building using co-association

Problems in generalized linear model selection and predictive evaluation for binary outcomes

Fuzzy Partitioning with FID3.1

An Ensemble of Classifiers using Dynamic Method on Ambiguous Data

Efficient Pruning Method for Ensemble Self-Generating Neural Networks

Instance-Based Learning with Genetically Derived Attribute Weights

Journal of Babylon University/Pure and Applied Sciences/ No.(2)/ Vol.(25): 2017

Evaluating the Replicability of Significance Tests for Comparing Learning Algorithms

THE discrete multi-valued neuron was presented by N.

An Empirical Study of Hoeffding Racing for Model Selection in k-nearest Neighbor Classification

Machine Learning in Biology

Rita McCue University of California, Santa Cruz 12/7/09

Data Mining: An experimental approach with WEKA on UCI Dataset

CS145: INTRODUCTION TO DATA MINING

INTRODUCTION to SAS STATISTICAL PACKAGE LAB 3

A Maximal Margin Classification Algorithm Based on Data Field

Experimental Approach for the Evaluation of Neural Network Classifier Algorithms

Adaptive Metric Nearest Neighbor Classification

Classifier Inspired Scaling for Training Set Selection

OPTIMUM COMPLEXITY NEURAL NETWORKS FOR ANOMALY DETECTION TASK

PCA-based Offline Handwritten Character Recognition System

Dataset Editing Techniques: A Comparative Study

Maximum Margin Binary Classifiers using Intrinsic and Penalty Graphs

Global Metric Learning by Gradient Descent

Create a SAS Program to create the following files from the PREC2 sas data set created in LAB2.

Comparative analysis of classifier algorithm in data mining Aikjot Kaur Narula#, Dr.Raman Maini*

Fuzzy Modeling using Vector Quantization with Supervised Learning

Nearest Cluster Classifier

Nearest Cluster Classifier

LEARNING WEIGHTS OF FUZZY RULES BY USING GRAVITATIONAL SEARCH ALGORITHM

Cost-Conscious Comparison of Supervised Learning Algorithms over Multiple Data Sets

An Empirical Comparison of Ensemble Methods Based on Classification Trees. Mounir Hamza and Denis Larocque. Department of Quantitative Methods

Intro to Artificial Intelligence

Distance Learning and Attribute Importance Analysis by Linear Regression on Idealized Distance Functions

Reihe Informatik 10/2001. Efficient Feature Subset Selection for Support Vector Machines. Matthias Heiler, Daniel Cremers, Christoph Schnörr


A Comparative Study of Selected Classification Algorithms of Data Mining

Application of Clustering as a Data Mining Tool in Bp systolic diastolic

Decision Jungles: Compact and Rich Models for Classification Supplementary Material

Data Mining and Knowledge Discovery. Data Mining and Knowledge Discovery

Value Difference Metrics for Continuously Valued Attributes

Class Strength Prediction Method for Associative Classification

ECG782: Multidimensional Digital Signal Processing

Practice EXAM: SPRING 2012 CS 6375 INSTRUCTOR: VIBHAV GOGATE

CHAPTER 3 RESEARCH METHODOLOGY

Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions

Fingerprint Based Gender Classification Using Block-Based DCT

Some Thoughts on Machine Learning Software Design

Assignment 1: CS Machine Learning

Transcription:

1 2 2 Paul Horton 3 2 One-Against-One OAO One-Against-All OAA UCI k-nn OAO SVM A proposal of hierarchization method based on data distribution for multi class classification Tatsuya Kusuda, 1 Shinya Watanabe, 2 Jianming Shi 2 and Paul Horton 3 This paper proposes a new hierarchical method for multi-class classification using binary classifiers. Unlike existing extension methods, such as One- Against-One (OAO) and One-Against-All (OAA), our proposed method makes a hierarchic structure of classification according to distributions of each class data as given training data. Thus, a more accurate multi-class classification can be expected than by existing methods. In particular, our proposed method is expected to be more effective for classes with few samples. Since a hierarchical structure of classification derived by our proposed method is formed on the basis of similarities among classes, relative features of each class can be inferable through the classification structure. In this paper, the effectiveness of our proposed method is discussed through some examples from UCI repository, based on comparison with that of k-nn and OAO. 1. SVM 1) 4) k k-nearest neighbor algorithm k-nn SVM 2 2 (Support Vector Machine, SVM) 1) 2 One-Against-One OAO One-Against-All OAA 5) 6) OAO OAA 5) 1 Graduate School of Information and Electronic Engineering, Muroran Institute of Technology 2 Department of Information and Electronic Engineering, Muroran Institute of Technology 3 National Institute of Advanced Industrial Science and Technology Computational Biology Research Center (AIST CBRC) 1

UCI 7) k-nn SVM OAO 2 3 UCI 2. 1 1 1 hierarchy hierarchy hierarchy classifier 3 1 classifier 2 3 classifier4 class 2 Conceptual diagram of proposal method classifier class 2 3 classifier5 class 3 1 1 3 2 3 4 2 1 3 2 3 4 3 3 4 2.1 2 Step1: Step2: 2.2 Step3: Step2 SVM Step4: Step3 Step2 Step1:. Step2: SVM Step3: OAO SVM Step2 2.2 2-means k-means k=2 2

2-means 2-means 8) 2 0 2 0,1,2 2 1 1 Step1: 2-means 2 Step2: Step1 Step3: α% Step4: N Gj Step5: C G Balance Accuracy Eval = β Balance + (1 β) Accuracy (1) Balance = G j=1 ( N Gj G C ) (2) Accuracy = Step6: Step1 Step5 (3) Step7: Step1 Step6 Step8: (1) 1 2 β 0.3 Step3 α α 3. UCI k-nn SVM OAO 3.1 UCI 7) 6 Iris,Wine,Heart Disease,Glass,Vowel,Car Evaluation 1 3.2 5 α β Balance (1) 2-means 2 Car Evaluation α 0.2 α 0.25 3.3 UCI 6 Iris Wine Heart Disease Glass Vowel Car Evaluation 1 Car Evaluation 0.25 3

1 The characteristics of the used data Dataset Number Features Classes of data Iris 150 4 3 Wine 178 13 3 Heart Disease 270 13 5 Glass 214 10 6 Vowel 528 10 11 Car Evaluation 1728 7 4 Parameter 2 Used parameters Determination of group parameter α Values 0.2 1 Weight of evaluation 0.3 formula parameter β Number of partitions 10 in cross validation Number of clustering 30 4 Glass The predictive accuracy of each class in Glass Id Number of data Weighted k-nn OAO Proposed method 1 70 81.43% 71.43% 84.29% 2 76 67.11% 76.32% 71.05% 3 17 0.00% 11.76% 17.65% 5 13 46.15% 53.85 % 46.15% 6 9 22.22% 22.22% 44.44% 7 29 82.76% 79.31% 86.21% 214 65.42% 66.36% 70.56% 3 The results of predictive accuracy Dataset Weighted k-nn OAO Proposed method Iris 95.33% 96.00% 97.33% Wine 93.26% 97.19% 97.75% Heart Disease 55.56% 55.22% 56.57% Glass 65.42% 66.36% 70.56% Vowel 92.02% 99.60% 95.04% Car Evaluation 88.60% 99.36% 99.07% 3.3.1 10 (10-fold Cross-Validation) k-nn SVM OAO SVM 3 3 3 Iris Wine Heart Disease Glass Vowel Car Evaluation OAO Vowel Car Evaluation OAO α Glass Glass 4 4 Glass 2 OAO 5 α The effects of α value for predictive accuracy Dataset Weighted OAO Proposed method k-nn α values 0.4 0.3 0.2 0.1 0.0 Iris 94.00% 94.67% 96.00% 95.33% 95.33% 96.00% 96.67% Wine 93.26% 97.19% 97.75% 97.19% 97.75% 98.31% 97.19% Heart Disease 54.88% 54.88% 54.88% 54.88% 55.22% 54.88% 54.88% Glass 65.42% 64.95% 64.49% 64.49 % 64.95% 67.76% 64.95% Vowel 92.02% 99.60% 95.45% 95.15% 95.05% 95.75% 99.60% Car Evaluation 82.81% 99.71% 99.42% 99.54% 99.71% 99.71% 99.71% 3 6 7 α α α α α 5 5 α α OAO 4

id 6 Glass Classes of Glass data Details of classes 1 building windows float processed 2 building windows non float processed 3 vehicle windows float processed 5 containers 6 tableware 7 headlamps 7 Glass The attributes of Glass data id Features name Details of features 1 RI refractive index 2 Na Sodium 3 Mg Magnesium 4 Al Aluminum 5 Si Silicon 6 K Potassium 7 Ca Calcium 8 Ba Barium 9 Fe Iron OAO 3.3.2 Glass Heart Disease Glass UCI 6 7 Glass 2 2 2 2 1 2 3 1 2 3 8 9 Heart Disease Heart Disease 3 3 2 id 9 2 Glass Hierarchy diagram of proposal technique(glass) Classes id 1 4 8 Heart Disease Classes of Heart Disease data Details of classes 0 person who hasn t heart disease person who has heart disease 1 4 is degree 9 Heart Disease The attributes of Heart Disease data Features id Features name Details of features 1 age age in years 2 sex sex (1 = male; 0 = female) 3 cp chest pain type 4 trestbps resting blood pressure 5 chol serum cholestoral in mg/dl 6 fbs fasting blood sugar 120 mg/dl 7 restecg resting electrocardiographic results 8 thalach maximum heart rate achieved 9 exang exercise induced angina 10 oldpeak ST depression induced by exercise relative to rest 11 slope the slope of the peak exercise ST segment 12 ca number of major vessels (0-3) colored by flourosopy 13 thal 3 = normal; 6 = fixed defect; 7 = reversable defect 5

3 Heart Disease 0.2 Hierarchy diagram of proposal technique(heart Disease 0.2) 4 (Heart Disease 0.4) Hierarchy diagram of proposal technique(heart Disease 0.4) 3 2 3 4 0 1 α 0.4 4 4 1 0 4 5 8 10 1 2 4. UCI k-nn OAO OAO 1) G. Nalbantov P.J.F.Groenen and J.C.Bioch. a majorization approach to linear support vector machines with different hinge errors. Advances in Data Analysis and Classification, Vol. 2, pp. 17-43, 2008. 2) J.A.K. Suykens and J. Vandewalle: Least squares support vector machine classifiers, Neural Netherlands, Vol. 9, No. 3, pp. 293-330, 6.1999. 3) Doumpos, M. Zopounidis, C. Golfinopoulou, V: Additive Support Vector Machines for Pattern Classification, IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, Vol. 37, No. 3, pp. 540-550, 2007. 4) Fukumizu, K: Special statistical properties of neural network learning. Proc, NOLTA 97, pp. 747-750, 1997. 5) Jonathan Milgram and Mohamed Cheriet and Robert Sabourin: One Against One or One Against All :Which One is Better for Handwriting Recognition with SVMs?, Ecole de Technologie Superieure, Montreal, Canada, 2006 6) Chih-Wei Hsu Chih-Jen Lin: A comparison of methods for multiclass support vector machines, Neural Networks, IEEE Transactions on, Vol. 13, No. 2, pp. 415-425, 2002. 7) C.L.Blake and C.J.Merz: UCI repository of machine learning databases, University of California, Department of Information and Computer Science, 1998, http://www.ics.uci.edu/mlrepository.html 8),,,,1988 9),,1994 10) Baldi P et al: Assessing the accuracy of prediction algorithms for classification, an overview, Bioinformatics, Vol. 16, No. 5, pp. 412-424, 2000 6