Feature weighting using particle swarm optimization for learning vector quantization classifier

Similar documents
Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India.

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

Improving Tree-Based Classification Rules Using a Particle Swarm Optimization

GENETIC ALGORITHM VERSUS PARTICLE SWARM OPTIMIZATION IN N-QUEEN PROBLEM

Feature Selection Algorithm with Discretization and PSO Search Methods for Continuous Attributes

Novel Initialisation and Updating Mechanisms in PSO for Feature Selection in Classification

International Journal of Digital Application & Contemporary research Website: (Volume 1, Issue 7, February 2013)

Modified Particle Swarm Optimization

IMPROVING THE PARTICLE SWARM OPTIMIZATION ALGORITHM USING THE SIMPLEX METHOD AT LATE STAGE

Figure (5) Kohonen Self-Organized Map

A MULTI-SWARM PARTICLE SWARM OPTIMIZATION WITH LOCAL SEARCH ON MULTI-ROBOT SEARCH SYSTEM

Traffic Signal Control Based On Fuzzy Artificial Neural Networks With Particle Swarm Optimization

Mobile Robot Path Planning in Static Environments using Particle Swarm Optimization

Hybrid Particle Swarm-Based-Simulated Annealing Optimization Techniques

A NEW APPROACH TO SOLVE ECONOMIC LOAD DISPATCH USING PARTICLE SWARM OPTIMIZATION

Comparing Classification Performances between Neural Networks and Particle Swarm Optimization for Traffic Sign Recognition

An improved PID neural network controller for long time delay systems using particle swarm optimization algorithm

Particle Swarm Optimization applied to Pattern Recognition

Simplifying Handwritten Characters Recognition Using a Particle Swarm Optimization Approach

Particle Swarm Optimization

Index Terms PSO, parallel computing, clustering, multiprocessor.

Conveyor Performance based on Motor DC 12 Volt Eg-530ad-2f using K-Means Clustering

Overlapping Swarm Intelligence for Training Artificial Neural Networks

Clustering of datasets using PSO-K-Means and PCA-K-means

OPTIMUM CAPACITY ALLOCATION OF DISTRIBUTED GENERATION UNITS USING PARALLEL PSO USING MESSAGE PASSING INTERFACE

Improving the accuracy of k-nearest neighbor using local mean based and distance weight


Hybrid PSO-SA algorithm for training a Neural Network for Classification

Optimization Model of K-Means Clustering Using Artificial Neural Networks to Handle Class Imbalance Problem

A Fast Wrapper Feature Subset Selection Method Based On Binary Particle Swarm Optimization

A Multiobjective Memetic Algorithm Based on Particle Swarm Optimization

CHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES

Discrete Particle Swarm Optimization With Local Search Strategy for Rule Classification

PARTICLE SWARM OPTIMIZATION FOR MULTIDIMENSIONAL CLUSTERING OF NATURAL LANGUAGE DATA

PSO algorithm for an optimal power controller in a microgrid

PARALLEL PARTICLE SWARM OPTIMIZATION IN DATA CLUSTERING

Particle swarm optimization for mobile network design

Reconfiguration Optimization for Loss Reduction in Distribution Networks using Hybrid PSO algorithm and Fuzzy logic

Design optimization method for Francis turbine

Adaptive Radiation Pattern Optimization for Antenna Arrays by Phase Perturbations using Particle Swarm Optimization

Handling Multi Objectives of with Multi Objective Dynamic Particle Swarm Optimization

Open Access Research on the Prediction Model of Material Cost Based on Data Mining

Prime Numbers Comparison using Sieve of Eratosthenes and Sieve of Sundaram Algorithm

Application of Improved Discrete Particle Swarm Optimization in Logistics Distribution Routing Problem

Initializing the Particle Swarm Optimizer Using the Nonlinear Simplex Method

Vigenere cipher algorithm modification by adopting RC6 key expansion and double encryption process

SIMULTANEOUS COMPUTATION OF MODEL ORDER AND PARAMETER ESTIMATION FOR ARX MODEL BASED ON MULTI- SWARM PARTICLE SWARM OPTIMIZATION

Particle Swarm Optimization Approach for Scheduling of Flexible Job Shops

PATTERN RECOGNITION USING NEURAL NETWORKS

Cell-to-switch assignment in. cellular networks. barebones particle swarm optimization

Object Recognition using Particle Swarm Optimization on Fourier Descriptors

Particle Swarm Optimization

Inertia Weight. v i = ωv i +φ 1 R(0,1)(p i x i )+φ 2 R(0,1)(p g x i ) The new velocity update equation:

PARTICLE SWARM OPTIMIZATION (PSO)

Particle Swarm Optimization Artificial Bee Colony Chain (PSOABCC): A Hybrid Meteahuristic Algorithm

Tracking Changing Extrema with Particle Swarm Optimizer

2009 International Conference of Soft Computing and Pattern Recognition

Particle Swarm Optimization Based Approach for Location Area Planning in Cellular Networks

A Comparative Analysis on the Performance of Particle Swarm Optimization and Artificial Immune Systems for Mathematical Test Functions.

A RANDOM SYNCHRONOUS-ASYNCHRONOUS PARTICLE SWARM OPTIMIZATION ALGORITHM WITH A NEW ITERATION STRATEGY

Particle Swarm Optimization Applied to Job Shop Scheduling

Three-Dimensional Off-Line Path Planning for Unmanned Aerial Vehicle Using Modified Particle Swarm Optimization

Convolutional Code Optimization for Various Constraint Lengths using PSO

LECTURE 16: SWARM INTELLIGENCE 2 / PARTICLE SWARM OPTIMIZATION 2

A Native Approach to Cell to Switch Assignment Using Firefly Algorithm

Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic Algorithm and Particle Swarm Optimization

KEYWORDS: Mobile Ad hoc Networks (MANETs), Swarm Intelligence, Particle Swarm Optimization (PSO), Multi Point Relay (MPR), Throughput.

A PSO-based Generic Classifier Design and Weka Implementation Study

Automatic differentiation based for particle swarm optimization steepest descent direction

A Study on Optimization Algorithms for Clustering Gene Expression Data

CHAPTER 6 COUNTER PROPAGATION NEURAL NETWORK IN GAIT RECOGNITION

IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS PART B: CYBERNETICS /$ IEEE

A Comparative Study of the Application of Swarm Intelligence in Kruppa-Based Camera Auto- Calibration

Algorithm for Classification

Kyrre Glette INF3490 Evolvable Hardware Cartesian Genetic Programming

Discrete Particle Swarm Optimization for Solving a Single to Multiple Destinations in Evacuation Planning

Small World Particle Swarm Optimizer for Global Optimization Problems

Feeder Reconfiguration Using Binary Coding Particle Swarm Optimization

Solving Economic Load Dispatch Problems in Power Systems using Genetic Algorithm and Particle Swarm Optimization

Measuring the power consumption of social media applications on a mobile device

Research Article Path Planning Using a Hybrid Evolutionary Algorithm Based on Tree Structure Encoding

Particle Swarm Optimization For N-Queens Problem

Estimating Missing Attribute Values Using Dynamically-Ordered Attribute Trees

Week 9 Computational Intelligence: Particle Swarm Optimization

Optimized Algorithm for Particle Swarm Optimization

Improved PSO for Feature Selection on High-Dimensional Datasets

A HYBRID ALGORITHM BASED ON PARTICLE SWARM OPTIMIZATION

Computer-Aided Diagnosis for Lung Diseases based on Artificial Intelligence: A Review to Comparison of Two- Ways: BP Training and PSO Optimization

Robust Descriptive Statistics Based PSO Algorithm for Image Segmentation

INTEGRATION OF INVENTORY CONTROL AND SCHEDULING USING BINARY PARTICLE SWARM OPTIMIZATION ALGORITHM

Kent Academic Repository

Seismic regionalization based on an artificial neural network

Classification Using Unstructured Rules and Ant Colony Optimization

Effect of the PSO Topologies on the Performance of the PSO-ELM

SOMSN: An Effective Self Organizing Map for Clustering of Social Networks

Artificial Neural Network Training by using Regrouping Particle Swarm Optimization

Nearby Search Indekos Based Android Using A Star (A*) Algorithm

LEARNING WEIGHTS OF FUZZY RULES BY USING GRAVITATIONAL SEARCH ALGORITHM

Constraints in Particle Swarm Optimization of Hidden Markov Models

Witold Pedrycz. University of Alberta Edmonton, Alberta, Canada

Transcription:

Journal of Physics: Conference Series PAPER OPEN ACCESS Feature weighting using particle swarm optimization for learning vector quantization classifier To cite this article: A Dongoran et al 2018 J. Phys.: Conf. Ser. 978 012032 View the article online for updates and enhancements. This content was downloaded from IP address 148.251.232.83 on 27/09/2018 at 21:51

Feature weighting using particle swarm optimization for learning vector quantization classifier A Dongoran* 1, S Rahmadani 1, M Zarlis 1 and Zakarias 1 1 Magister Teknik Informatika, Fasilkom - TI, Universitas Sumatera Utara (USU), Medan - Indonesia *Awaluddin.dongoran@gmail.com Abstract. This paper discusses and proposes a method of feature weighting in classification assignments on competitive learning artificial neural network LVQ. The weighting feature method is the search for the weight of an attribute using the PSO so as to give effect to the resulting output. This method is then applied to the LVQ-Classifier and tested on the 3 datasets obtained from the UCI Machine Learning repository. Then an accuracy analysis will be generated by two approaches. The first approach using LVQ1, referred to as LVQ-Classifier and the second approach referred to as PSOFW-LVQ, is a proposed model. The result shows that the PSO algorithm is capable of finding attribute weights that increase LVQ-classifier accuracy. 1. Introduction Classification is a process of learning by mapping a data into one or more predefined classes based on information discribed by its own features [6]. Learning that occurs in the classification process is also called an inductive-learning algorithm [1]. The purpose of the classification is to create a model that will be used to predict the value of the input data that is not yet known class. Simply split the classification process usually divide the sample data into two parts [2]. First referred to as training data used to train the data to get the expected classification model. The two are referred to as test data. This test data will be tested to determine the level of accuracy in predicting the value of new input data, or can also known error rate predictions. A statistical approach can also be used to generate a classification model. In addition to statistical approaches, there are many more approaches used to generate classification models, such as decision tree, artificial neural network, genetic algorithm, evolutionary computing, swarm optimization, etc. In this paper we will use competitive learning method based on competition on LVQ competitive network, and using particle swarm optimization (PSO) method to weighting each attribute of sample data. Both of these approaches will be compared with each other to find out the performance of both approaches in predicting the value of new mapped input data. The weighting method feature improves performance in this case increases accuracy or decreases error rate [5]. The weighting feature is part of the selection feature. The selection feature is a fairly complex job that can affect the performance of classification. Therefore approaches to feature selection continue to be developed. The most popular current approach is based on Evolutionary Computation (EC), Particle Swarm Optimization (PSO) [7] [8] based Swarm Intelligence. PSO is a simple and fast Content from this work may be used under the terms of the Creative Commons Attribution 3.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI. Published under licence by Ltd 1

computing technique that produces convergent values. So PSO is an effective technique used for feature selection tasks [10]. 2. Learning Vector Quantization (LVQ) Learning vector quantization (LVQ) is a method that can be used to generate a classification model of trained sample data. LVQ is part of a supervised neural network method [9]. LVQ architecture is similar to Kohonen's SOM (Self Organizing Map) developed by Prof. Teuvo Kohonen in 1982. Simply put, competitive layer on artificial neural networks have a single output layer, and each neuron is fully connected to the input node [1]. Layer of inputs x 1 Single layer of output nodes Competitive Outputs y 1 x 2 y 2 x n y k Figure 1. Architecture Simple LVQ 2.1. LVQ Algorithm The value of each attribute (information) training data is used to form a weight vector (prototype). The weight vector then iteratively becomes a model that will be used to classify the new sample data whose class is not yet known. The first stage of the LVQ algorithm is to determine the initial weights in the classification model search process. The initial weights used are derived from the training data samples. LVQ applies a winner-take-all strategy [5], choosing a winning vector as a winner. The winning weight vector is the vector that has the smallest distance or in other words, this weight has a resemblance to the input vector. Then the winning weights vector are updated or adjusted. While other weight vectors (which are not winners) are not updated. The weight vector update process is performed repeatedly until the termination conditions are met. The size of the similarity between the weight vector and the input vector is determined using the distance (Euclidean) formula in equation 1 below. dist(x, w) = n i=1 (x i w i ) 2 (1) Where n is the number of attributes, x is the input vector, and w is the weight vector. In pattern recognition, the minimum Euclidean distance between the input vectors and the vector weights is the winner. This winning vector is then updated. If the value of input vector x and weight vector w (winner) are in the same class, then the weight vector w is updated using equation 2. w(t + 1) = w(t) + α(x(t) w(t)) (2) And vice versa, if the input vector x and weight vector w (winner) are in different classes, and then the weight vector w is updated using equation 3. Where α is called the learning rate. 2

w(t + 1) = w(t) α(x(t) w(t)) (3) The steps of LVQ learning process to generate the classification model, as follows: 1. Input: Input data, vector codebook, learning rate, maximum iteration 2. Initialization vector weights (taken from data input) 3. For each data input, calculate the Euclidean distance between each data input x and the weight vector w using equation 1 4. For each weighted vector w, select the smallest Euclidean distance as the winner. 5. Update the winning weight vector w: If class w is equal to class x, update w using equation 2. If class w is not the same as class x, update w uses equation 3. 6. If the termination condition is met, print the weighted vector w as the resulting classification model, otherwise return to step 4 3. Particle Swarm Optimization (PSO) The swarm particle is a population-based optimization technique developed by James Kennedy and Russ Eberhart in 1995 [7] [8] inspired by the behavior of bird flocks. Although Kennedy and Eberhart started modeling with flocks, it turned out that in its development, the algorithm model is more similar to a model of swarming bees with collective intelligence [3]. 3.1. PSO Algorithm The particle swarm-based optimization begins by initializing the positions of the particles randomly. Then each particle calculated its fitness value based on the evaluation function. This method will continue to iterate until the expected condition is reached (optimal condition). At each iteration, each particle is affected by two values [5]. The first is the best value each particle has achieved, called the pbest. The other value is the best overall value possessed by particles in the population, referred to as gbest. After obtaining the two best values, the particles renewed the velocity and their respective positions with equations 4 and 5. v i (t) = v i (t 1) + c 1 r 1 (x pi x i ) + c 2 r 2 (x g x i ) (4) x i (t) = x i (t 1) + v i (t) (5) Where v represents velocity, x is the particle position, c is the learning rate and r is a random number U (0,1). v i (t 1) is also called inertia term [11] that forces particles to move in the same direction as the previous movement. c 1 r 1 (x pi x i ) Is referred to as the cognitive term [11] that forces the particle 3

back to its best position. c 2 r 2 (x g x i ) is referred to as social term [11] which forces particles to move to the best position of all particles. The steps for the particle swarm-based optimization procedure are as follows: 1. Generate the initial population of the particles 2. Calculate the fitness value, if better than the fitness value during the run, set that value as pbest 3. Choose the particle with the best fitness value in the population, set that value as gbest 4. For each particles: a. Update velocity v using equation 4 b. Update position x using equation 5 5. Stop if expected conditions are reached. If not, returned to step 2 4. Experiment In this paper used two methods, namely LVQ and PSO. PSO as a weight seeker on feature (attribute) sample data, while LVQ is used to find the classification model (classifier). In the PSO algorithm, the datainput feature is used to search for the weight value, then the weight value is used as the LVQ input parameter to generate the classification model through the training process. The classification model is then validated in the testing process.tthe performance level is then calculated for looking at the accuracy and error rate. The data used for training and testing is obtained from sample dataset using Rotation (n-fold cross validation). 4.1. Model PSO to find Feature Weighting The particle swarm-based optimization begins with the initialization of the population of the particles by randomize way. The particles represent the sample dataset s attributes. The value of each attribute is randomized (from the minimum to the maximum value) to set the initial position of the particle. While the initial velocity of the particles is set to zero. The particle density must not exceed the maximum value of each attribute of the sample dataset. After initializing the value, each particle is then evaluated by an evaluation function to find its pbest value [5]. From the value obtained pbest, select one value with the best value as the gbest value [5]. Then update the velocity and position of each particle using equations 4 and 5. The fitness value evaluation process, density and position update, is done iteratively until the expected termination conditions are achieved. The value of the learning rate is updated iteratively to see the resulting convergence level. Convergence is met if the evaluation function tends to improve and stagnate, and the position of the particles has not changed significantly. The output of this PSO algorithm is the weighted value of each attribute of dataset. 4.2. Model LVQ to find Model Classifier The weight of the attributes obtained from the PSO algorithm is then applied to the LVQ model. The weight of the attribute a j, where j is the j s attribute, is inserted into the euclidean distance formula, as shown in equation 6. dist(x, w) = n (a j (x i w k )) 2 i=1 (6) Where x is the input data, w k is vector weight for each class k, a j is the feature weight generated by the PSO on each attribute j. While the formula for the winner vector weighted update is modified due to the weight value of the resulting attribute. If the input x and the weight vector w are in the same class then the weight w is updated using equation 7, otherwise if x and w are not in the same class, the weight w is updated using equation 8. 4

w (t+1) = a w (t) + α (a (x (t) w (t) )) (7) w (t+1) = a w (t) + α (a (x (t) w (t) )) (8) Where a is the attribute s weight, α is a learning rate. 5. Result The dataset that used to train and test the above experiments was obtained from UCI Machine Learning Repository. The dataset has been classified before, each instance already has a certain class label. Table 1 shows a description of the 3 datasets with property A representing an attribute, I denoting the number of instances, C denoting the number of classes. Table 2 shows the optimal attribute weights generated by the PSO algorithm of each dataset. Table 3 shows the data validation results consisting of accurate predictive value, and error rate. Table 1. Description of Datasets. Attribute. (A) Class. (C) Instance. (I) Iris 4 3 150 Glass 9 7 214 Wine 13 3 178 Table 2. Attribute Weight of Datasets. Iris Glass Wine Atrr Weight Attr Weight Attr Weight Attr Weight A1 0.981 A1 1 A1 0.249 A10 0.336 A2 0.745 A2 0.640 A2 0 A11 1 A3 0.676 A3 1 A3 0.398 A12 0 A4 0.447 A4 0.540 A4 0 A13 0 A5 0.290 A5 0 A6 0.984 A6 0.978 A7 1 A7 1 A8 1 A8 1 A9 0 A9 0 5

Table 3. Accuracy of Prediction. Dataset Accuracy Classification Error LVQ-Classifier PSOFW-LVQ LVQ-Classifier PSOFW-LVQ Iris 96.00% 97.33% 4.00% 3.27% Glass 72.55% 77.06% 27.45% 22.94% Wine 90.98% 98.33% 9.02% 1.67% 6. Summary In this paper, to form an optimal classification model used two methods namely, LVQ and PSO. The result of data processing and validation shows that there is an increase of classifier accuracy performance if it is weighted on attribute of each dataset. This improvement in accuracy shows that the PSO algorithm is capable of finding attribute weights that increase LVQ-classifier accuracy, although LVQ-Classifier itself has produced a fairly good classification model. Experiments conducted on this papaer is devoted to continuous values, future will be tried for non-continue values. References [1] Mehmed K 2003 Data Mining: Concept, Models, Methods, And Algorithms IEEE Press (445 Hoes Lane, Piscataway, NJ 08854) [2] Witten I H, Frank E and Hall M A 2011 Data Mining: Practical Machine Learning Tools and Techniques Morgan Kaufmann Publisher. Elsevier.Inc (Burlington, MA 01803, USA) [3] Millonas M 1994 Swarms, Phase Transition, and Collective Inteligence [4] Cardic C 1993 Proc.Int.Conf. on Machine Learning [5] Brownlee J 2011 Clever Algorithms: Nature-Inspired programming Recipes Creative Commons Attribution-Noncommercial-Share Alike 2.5 (Autralia License) ISBN: 978-1-4467-8506-5 [6] Xue B, Zhang M and Browne W N 2012 Particle Swarm Optimization for Feature Selection in Classification: A Multi Objective Approach IEEE Transaction On Cybernetics Journal. IEEE Press [7] Kennedy J and Eberhart R 1995 Proc. IEEE Int.Conf. Neural Network vol. 4 [8] Shi Y and Eberhart R 1998 Proc. IEEE Int.CEC pp. 69-73. [9] Kohonen T 1988 Self-Organization and Associative Memory Springer-Verlag (New York) [10] Liu Y, Wang G, Chen H and Dong H 2011 J.Bionic Eng.,vol.8 2 pp.191-200 [11] Alexandre S and Leandro N d C 2013 A Constructive Data Classification Version of the Particle Swarm Optimization Algorithm Hindawi Publishing Corporation Mathematical Problems in Engineering 6