A Modified Fuzzy Min-Max Neural Network and Its Application to Fault Classification

Similar documents
Modified Fuzzy Hyperline Segment Neural Network for Pattern Classification and Recognition

Introduction to Artificial Intelligence

A Fuzzy C-means Clustering Algorithm Based on Pseudo-nearest-neighbor Intervals for Incomplete Data

APPLICATION OF THE FUZZY MIN-MAX NEURAL NETWORK CLASSIFIER TO PROBLEMS WITH CONTINUOUS AND DISCRETE ATTRIBUTES

An Empirical Comparison of Ensemble Methods Based on Classification Trees. Mounir Hamza and Denis Larocque. Department of Quantitative Methods

T-S Neural Network Model Identification of Ultra-Supercritical Units for Superheater Based on Improved FCM

LABVIEW APPLICATION: ENERGY LABORATORY UPGRADE

Unsupervised Learning : Clustering

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION

Associative Cellular Learning Automata and its Applications

CHAPTER 3 MODELING OF DEAERATOR AND SIMULATION OF FAULTS

MODELING FOR RESIDUAL STRESS, SURFACE ROUGHNESS AND TOOL WEAR USING AN ADAPTIVE NEURO FUZZY INFERENCE SYSTEM

Cluster Analysis: Agglomerate Hierarchical Clustering

Pattern Classification Algorithms for Face Recognition

EVALUATION OF CONVENTIONAL DIGITAL CAMERA SCENES FOR THEMATIC INFORMATION EXTRACTION ABSTRACT

INF4820 Algorithms for AI and NLP. Evaluating Classifiers Clustering

Unsupervised Learning

K-means clustering Based in part on slides from textbook, slides of Susan Holmes. December 2, Statistics 202: Data Mining.

Spatial Information Based Image Classification Using Support Vector Machine

List of forms CLAP/WGP by subject /68/UE

The k-means Algorithm and Genetic Algorithm

Heart Disease Detection using EKSTRAP Clustering with Statistical and Distance based Classifiers

EE 589 INTRODUCTION TO ARTIFICIAL NETWORK REPORT OF THE TERM PROJECT REAL TIME ODOR RECOGNATION SYSTEM FATMA ÖZYURT SANCAR

Object Recognition Using Reflex Fuzzy Min-Max Neural Network with Floating Neurons

Modeling of Compressors and Expansion Devices With Two-Phase Refrigerant Inlet Conditions

Using a genetic algorithm for editing k-nearest neighbor classifiers

A comparative study of local classifiers based on clustering techniques and one-layer neural networks

Clustering & Classification (chapter 15)

FUZZY KERNEL K-MEDOIDS ALGORITHM FOR MULTICLASS MULTIDIMENSIONAL DATA CLASSIFICATION

COMP33111: Tutorial and lab exercise 7

Early tube leak detection system for steam boiler at KEV power plant

Designing Ultra-Efficient Building Cooling Systems

University of Florida CISE department Gator Engineering. Clustering Part 5

Data Mining Clustering

An Empirical Study of Hoeffding Racing for Model Selection in k-nearest Neighbor Classification

CHAPTER 3 FUZZY RULE BASED MODEL FOR FAULT DIAGNOSIS

Advances in Military Technology Vol. 8, No. 2, December Applied Sensor Fault Detection, Identification and Data Reconstruction

Automatic Generation of Fuzzy Classification Rules Using Granulation-Based Adaptive Clustering

Figure (5) Kohonen Self-Organized Map

Gene Clustering & Classification

International Journal of Research in Advent Technology, Vol.7, No.3, March 2019 E-ISSN: Available online at

AN IMPROVED MULTI-SOM ALGORITHM

A REVIEW ON VARIOUS APPROACHES OF CLUSTERING IN DATA MINING

Final Report - Smart and Fast Sorting

Machine Learning - Clustering. CS102 Fall 2017

Basic Data Mining Technique

Review on Various Clustering Methods for the Image Data

CHAPTER 4 FUZZY LOGIC, K-MEANS, FUZZY C-MEANS AND BAYESIAN METHODS

A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation. Kwanyong Lee 1 and Hyeyoung Park 2

Classification Algorithms in Data Mining

Case-Based Reasoning. CS 188: Artificial Intelligence Fall Nearest-Neighbor Classification. Parametric / Non-parametric.

CS 188: Artificial Intelligence Fall 2008

CSE 158. Web Mining and Recommender Systems. Midterm recap

FEATURE EXTRACTION USING FUZZY RULE BASED SYSTEM

A Comparative Study of Conventional and Neural Network Classification of Multispectral Data

INF4820, Algorithms for AI and NLP: Evaluating Classifiers Clustering

Analysis of an airfoil

Land Cover Classification Techniques

Cluster Validity Classification Approaches Based on Geometric Probability and Application in the Classification of Remotely Sensed Images

Data Mining: STATISTICA

Online Adaptive Hierarchical Clustering in a Decision Tree Framework

Multi prototype fuzzy pattern matching for handwritten character recognition

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte

Hybrid Models Using Unsupervised Clustering for Prediction of Customer Churn

6. Dicretization methods 6.1 The purpose of discretization

A STUDY OF SOME DATA MINING CLASSIFICATION TECHNIQUES

FEATURE EXTRACTION TECHNIQUES USING SUPPORT VECTOR MACHINES IN DISEASE PREDICTION

4. Feedforward neural networks. 4.1 Feedforward neural network structure

Application of genetic algorithms and Kohonen networks to cluster analysis

Predicting Bias in Machine Learned Classifiers Using Clustering

Semi-Supervised Clustering with Partial Background Information

Clustering. Supervised vs. Unsupervised Learning

Applying Supervised Learning

Function Algorithms: Linear Regression, Logistic Regression

Adaptive Resolution Min-Max Classifiers

LEARNING WEIGHTS OF FUZZY RULES BY USING GRAVITATIONAL SEARCH ALGORITHM

Data Mining. Part 2. Data Understanding and Preparation. 2.4 Data Transformation. Spring Instructor: Dr. Masoud Yaghini. Data Transformation

MODELLING DOCUMENT CATEGORIES BY EVOLUTIONARY LEARNING OF TEXT CENTROIDS

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

CHAPTER 4 AN IMPROVED INITIALIZATION METHOD FOR FUZZY C-MEANS CLUSTERING USING DENSITY BASED APPROACH

FX2-CHILLER. Digital Control. Operations Manual

Supervised vs. Unsupervised Learning

KTH ROYAL INSTITUTE OF TECHNOLOGY. Lecture 14 Machine Learning. K-means, knn

Properties of learning of a Fuzzy ART Variant

Unsupervised Learning

Discriminate Analysis

ARTICLE; BIOINFORMATICS Clustering performance comparison using K-means and expectation maximization algorithms

2009 E09PS E09PS E09PS E09PS E09PS E09PS38 IEEE 2009 E09PS39 E09PS40 E09PS41 E09PS42 E09PS43 IEEE 2008 E09PS44

High throughput Data Analysis 2. Cluster Analysis

Jarek Szlichta

Validation of the Control Quality of Characteristic Field Based Fuzzy Controllers

Evolutionary Instance Selection Algorithm based on Takagi-Sugeno Fuzzy Model

Experimental Study on Fault Detection Algorithm Using Regression Method for Plural Indoor Units Faults of Multi-Heat Pump System under Heating Mode

Mass Classification Method in Mammogram Using Fuzzy K-Nearest Neighbour Equality

Outline. Prepare the data Classification and regression Clustering Association rules Graphic user interface

K-Nearest-Neighbours with a Novel Similarity Measure for Intrusion Detection

Machine Learning using MapReduce

Machine Learning (CSMML16) (Autumn term, ) Xia Hong

Constructive Feedforward ART Clustering Networks Part I

3 Feature Selection & Feature Extraction

Transcription:

A Modified Fuzzy Min-Max Neural Network and Its Application to Fault Classification Anas M. Quteishat and Chee Peng Lim School of Electrical & Electronic Engineering University of Science Malaysia

Abstract The objectives of this paper are: To improve the Fuzzy Min-Max (FMM) classification performance in situations when large hyperboxes are formed by the network. The Euclidean distance is computed after network training, and both the membership value of the hyperbox fuzzy sets and the Euclidean distance are used for classification. To assess the effectiveness of the modified FMM network. Benchmark pattern classification problems are first used, and the results from different methods are compared. In addition, a fault classification problem with real sensor measurements collected from a power generation plant is used to evaluate the applicability of the modified FMM network.

Introduction There are two types of learning classifiers: supervised and unsupervised learning, which differ in the way they are trained. The FMM neural network is a pattern classification system that can be used for tackling clustering (unsupervised) or classification (supervised) problems. In this paper FMM is used as a supervised classification system.

Introduction (contd.) FMM is constructed using hyperbox fuzzy sets, each of which is an n-dimensional box defined by a set of minimum and maximum points. Each input pattern is classified based on the degree of membership to the corresponding boxes. A smaller hyperbox size means that the hyperbox can contain only a smaller number of patterns, which will increase the network complexity, but yet gives high accuracy. A larger hyperbox size means that the hyperbox can contain a larger number of patterns, and will decrease the network complexity, but this leads to low classification accuracy.

FMM Neural Network The FMM classification network is formed using hyperbox fuzzy sets. A hyperbox defines a region of the n-dimensional pattern space that has patterns with full class membership. The hyperbox is completely defined by its minimum and maximum points. min point max point Fig. 1. A min-max hyperbox

The definition of each hyperbox fuzzy set B j is { XV,, W, f ( XV, W )} B, n j = j j j j X I (1) Where, X is the input, I n is a unit cube with n dimensions and V j and W j are the min and max points, respectively The size of the hyperboxes θ can have a value between 0 and 1. A small value of θ will produce small-size hyperboxes, and vice versa.

The input patterns are classified depending on how much they are contained by a hyperbox, this is measured using the membership value of a pattern using: b j n 1 ( Ah ) = [max γ 2n i= 1 ( 0,1 max ( 0, min ( 1, a w ) ( 0,1 max ( 0, min ( 1, v a ))] + max γ (2) A n ( a 1, a 2 a ) I ( ) =,..., where, h h h hn is the h th input pattern, V j = v j1, v j 2,..., v jn is the min point for B j, W j = ( w j1, w j2,..., w jn ) is the max point for B j, and γ is the sensitivity parameter. hi ji ji hi

Modifications of the FMM Network The modification done on the FMM neural network was done on the prediction stage. The original learning FMM procedure was not touched. In this paper we propose two methods for prediction using: 1. Euclidean distance 2. Both Euclidean distance and membership value

1) Prediction using the Euclidean distance This prediction method is based on the Euclidean distance between the input pattern and the centroid of the hyperbox. In addition to the min and max points, the centroid of patterns falling in each hyperbox is computed, as follows: C ji a hi C ji = C ji + (3) N j where C ji is the centroid of the j th hyperbox in the i th dimension, and N j is the number of patterns included in the j th hyperbox.

The Euclidean distance between the centroid and the input pattern is calculated using: E jh = n ( C ji a hi ) i= 1 2 (4) Where, E jh is the Euclidean distance between j th hyperbox and the h th input pattern. In the classification process, the hyperbox with the smallest Euclidean distance to the input pattern is selected as the winner, and the pattern is so classified that it belongs to this hyperbox.

Figure 2 shows the classification process of a twodimensional input pattern using the described method. Centroid of hyperbox 2 C 2 w 2 E 2 E 1 w 1 Centroid of hyperbox 1 C 1 Input v 2 Class 2 v 1 Class 1 Fig. 2. The classification process of an input pattern using the Euclidean distance

2) Prediction using both the membership function and Euclidean distance When θ is large, hyperbox sizes are large, as consequent when calculating the membership value more than one hyperbox will have very large membership value sometimes even unity value. To solve this problem we propose to use both the membership value and Euclidean distance for classification. The hyperboxes with the highest membership value are selected and then the Euclidean distance between the centroid of these boxes and the input pattern is calculated. The hyperbox with the smallest distance is used to classify the input pattern.

Experiments and results: The proposed methods were tested using 4 data sets: A) Bench mark data sets: 1) PID data set 2) IRIS data set B) Fault diagnosis data sets: 1) The heat transfer conditions. 2) The tube blockage conditions.

A) Bench mark data sets: 1) PID data set Pima Indian Diabetes (PID) data set consists of 768 samples in a two class problem. The set was divided into two sets, 75% for training and 25% for testing. The experiment was done 5 times and the average of the results is shown in Figure 3. A comparison between the results obtained by our method and other methods based on the same experimental criteria is as shown in Table 1.

Testing accuracy (%) 100 90 80 70 60 50 40 30 A B C 0.01 0.03 0.05 0.07 0.09 0.15 0.25 0.35 0.45 0.55 0.65 0.75 0.85 0.95 theta Fig. 3. The testing accuracy rates of the PID problem. Curve A shows the accuracy rate using the membership value only; curve B shows the accuracy rate using both the membership value and Euclidean distance; curve C shows the accuracy rate using the Euclidean distance only.

Table 1. Classification accuracy from various methods for the PID data set methods Accuracy (%) LDA 77.5 C4.5 73.0 CART 72.8 K-NN 71.9 Curve A (best result) 76.56 Curve B (best result) 72.4 Curve C (best result) 74.9

A) Bench mark data sets: 2) IRIS data set The IRIS data set consists 150 samples in a 3 class classification problem. The data set was divided into two sets; training set which consisted of 80% of each class, and a test set with the remaining samples. The experiment was conducted for the data set, and the results of the proposed methods along with the original FMM are shown in Figure 4. Table 2 shows the maximum accuracy of various methods in comparison with the proposed methods (all the experiments are conducted using the same training and test data sets).

96 Testing Accuracy % 94 92 90 88 86 84 A B C 82 0.01 0.03 0.05 0.07 0.09 0.15 0.25 0.35 0.45 0.55 0.65 0.75 0.85 0.95 Theta Fig. 4. The testing accuracy rates of the IRIS data set. Curve A shows the accuracy rate using the membership value only; curve B shows the accuracy rate using both the membership value and Euclidean distance; curve C shows the accuracy rate using the Euclidean distance only.

Table 2. Percentage error for various methods for the IRIS data set methods Accuracy (%) C4.5 91.60 OC1 93.89 LMDT 95.45 LVQ 92.55 Curve A (best result) 94.00 Curve B (best result) 93.33 Curve C (best result) 94.00

B) Fault Classification A fault detection and classification system predicts failures, and when a failure occurs, it identifies the reason(s) of failures. In this study, we investigate the applicability of modified FMM using a set of sensor measurements collected from a power generation plant in Malaysia. The system under consideration is a circulating water (CW) system, as shown in Figure 5

Low Pressure Turbines Steam To Sea Condenser Primary Bar Screen CW Pumps Common Discharge Header Strainer Seawater Condensate (Reused in steam cycle process) Fig. 5. The circulating water system

A data set of 2439 samples was collected. Each data sample consisted of 12 features comprising of temperature and pressure measurements at various inlet and outlet points of the condenser, as well as other important parameters. Two case studies were conducted: 1. The heat transfer conditions. 2. The tube blockage conditions.

B) Fault Classification 1) Heat Transfer Conditions The heat transfer conditions were classified into two categories: efficient or inefficient. From the data set, 1224 data samples (50.18%) that showed inefficient heat transfer condition, whereas 1215 data samples (49.82%) showed efficient heat transfer condition in the condenser. The data set (excluding one sample) was divided into two equal sets, each containing 1219 samples, one for training and the other for testing. Both data sets contained 50% of the data samples belonging to each class.

Figure 6 shows the testing accuracy of the proposed methods along with original FMM. Table 3 shows the testing accuracy along with the number of hyperboxes used in the classification.

Testing accuracy (%) 100 95 90 85 80 75 70 A B C 0.01 0.03 0.05 0.07 0.09 0.15 0.25 0.35 0.45 0.55 0.65 0.75 0.85 0.95 theta Fig. 6. The testing accuracy rates of the Heat Transfer Conditions. Curves A, B, and C show the testing accuracy rates using the membership value only, both the membership value and Euclidean distance, and the Euclidean distance only, respectively.

Table 3. Testing accuracy for the heat transfer data set Theta θ Membership value (%) Euclidean distance (%) Membership value and Euclidean distance (%) Number of hyperboxes 0.5 90.9 92.82 93.36 226 0.55 90.53 85.56 92.03 187 0.6 88.56 77.32 89.94 101 0.65 90.53 72.05 90.63 81 0.7 90.41 79.19 92.57 38 0.75 87.08 90.43 92.23 2 0.8 87.08 90.43 92.23 2 0.85 87.08 90.43 92.23 2 0.9 87.08 90.43 92.23 2 0.95 87.08 90.43 92.23 2

B) Fault Classification 2) Tube Blockage Conditions In this experiment, the objective was to predict the occurrence of tube blockage in the CW system. The conditions of the condenser tubes were categorized into two classes: significant blockage and insignificant blockage. The data set used in the previous experiment was again employed. A total of 1313 samples (53.83%) showed significant blockage and the remaining samples showed insignificant blockage in the condenser tubes. The data samples were divided into two sets for training and testing. Again, the data set (excluding one sample) was divided into two equal sets, each containing 1219 samples, one for training and the other for testing.

Figure 7 shows the testing accuracy of the proposed methods along with original FMM. Table 4 shows the testing accuracy along with the number of hyperboxes used in the classification.

testing accuracy (%) 105 100 95 90 85 80 75 A B C 0.01 0.03 0.05 0.07 0.09 0.15 0.25 0.35 0.45 0.55 0.65 0.75 0.85 0.95 theta Fig. 7. The testing accuracy rates of the Tube Blockage Conditions. Curves A, B, and C show the testing accuracy rates using the membership value only, both the membership value and Euclidean distance, and the Euclidean distance only, respectively.

Table 5. Testing accuracy for the block detection data set Theta θ Membership value (%) Euclidean distance (%) Membership value and Euclidean distance (%) Number of hyperboxes 0.5 99.88 96.61 100.00 264 0.55 100 85.56 99.68 155 0.6 100 86.79 95.28 78 0.65 91.88 83.39 93.75 62 0.7 92 85.85 92.13 81 0.75 91.88 85.36 93.01 21 0.8 92 86.35 89.25 42 0.85 92 88.98 96.41 6 0.9 92 93.75 96.75 2 0.95 92 93.75 96.75 2

Conclusions The results obtained reveal the usefulness of the proposed modifications in improving the performance of FMM when large hyperboxes are formed by the network.