MODEL FUZZY K-NEAREST NEIGHBOR WITH LOCAL MEAN FOR PATTERN RECOGNITION
|
|
- Meryl Miles
- 5 years ago
- Views:
Transcription
1 International Journal of Computer Engineering & Technology (IJCET) Volume 9, Issue 2, March-April 2018, pp , Article ID: IJCET_09_02_017 Available online at Journal Impact Factor (2016): (Calculated by GISI) ISSN Print: and ISSN Online: IAEME Publication MODEL FUZZY K-NEAREST NEIGHBOR WITH LOCAL MEAN FOR PATTERN RECOGNITION Hafizh Al-Kautsar Aidilof, Muhammad Zarlis, Syahril Efendi Department Computer Science, Faculty of Computer Science and Information Technology, Universitas Sumatera Utara, Medan, Indonesia ABSTRACT K-Nearest Neighbor is one of the top 10 algorithm in data mining (Wu, 2009). Based on its development, K-Nearest Neighbor is combined with Fuzzy's approach. Fuzzy K-Nearest Neighbor located as membership degrees - except euclidean distance - as a feature of the data attachment to the target class so that Fuzzy KNN is known to improve the classification results. Except for adding Fuzzy, K-Nearest Neighbor is also modified at the class determination stage with Local Mean rules. At Local Mean KNN, the value of the data vector's test were calculated in each target class so that the euclidian distance was not calculated between the data but it was also between the target classes. In this study, we divide the local mean vector of LMKNN by the degree of membership for each class produced by Fuzzy K-Nearest Neighbor to obtain a smaller value vector. This will affect the more obvious range of values of the trend of a data to a class than other class. The test was performed using Iris dataset with k taken as many as 3 nearest neighbors in each target class. Accuracy results obtained with data testing in each class are 93.3%, 86.6% and 100%, so the overall average is 93.3%. Key words: FKNN, LMKNN, Pattern Recognition. Cite this Article: Hafizh Al-Kautsar Aidilof, Muhammad Zarlis, Syahril Efendi, Model Fuzzy K-Nearest Neighbor with Local Mean for Pattern Recognition. International Journal of Computer Engineering and Technology, 9(2), 2018, pp INTRODUCTION K-nearest neighbor is known as one of the powerful data mining algorithms for solving classification problems (Wu, 2009). Not only is the problem of classification, k-nearest neighbor is also widely applied in pattern recognition and categorization of texts (Bhatia & Vandana, 2010; Jabbar, et al. 2013; Sánchez, et al., 2014). Some of the advantages of k- nearest neighbor are very nonlinear, fast, simple and easy to understand and apply (Wang, et al. 2007; GarcíaPedrajas & Ortiz-Boyer, 2009; Pan, et al., 2017; Ougiaroglou & Evangelidis, 2012; Song, et al., 2017). In addition to some advantages, k nearest neighbor certainly has its own weaknesses. Some weaknesses of k-nearest neighbor must use all training data to editor@iaeme.com
2 Model Fuzzy K-Nearest Neighbor with Local Mean for Pattern Recognition identify or classify, susceptible to data with high dimensionality and variable ranges, slow computing and inability to handle missing values (Rosyid, et al., 2013; Raikwal, 2012 ). Various studies were conducted to perfect the k-nearest neighbor. One is by applying fuzzy logic to determine the degree of membership of a test data against the target class. In addition, efforts to maximize the results are also done by modifying the way of determining the target class as the final stage of classification or pattern recognition from initially using the vote majority to local mean. Target class determination using a vote majority approach determines the class of a new data or test data based on the amount of data that dominates one of the classes of the class k taken and ignores the similarity of traits or patterns between data residing in different classes. In contrast to the local mean approach where the distance between test data and train data on different target classes is also taken into account so that the process will be more fair. 2. THEORETICAL BASIS 2.1. Fuzzy K-Nearest Neighbor Zbancioc (2012) quotes from Keller (-) wrote that the motivation using Fuzzy K-Nearest Neighbor is all vectors of classic K-Nearest Neighbor has a degree equivalent in clustering process so there is no measure of the level of the strength of the data against a target class. With the step fuzzyfication the data will be information describing the importance of the power of each vector in a final decision. Vector high value on the degree of fuzzy membership will have a value of more weight in the determination of the target class. The algorithm Fuzzy K-Nearest Neighbor described as follows (Beyan, 2014) : Determine the value of data training and data testing Perform data normalization if the range is too far Calculate the distance between the test data to training data with Euclidean equation Sort range from the smallest value to the largest value Retrieving as much as the value of k Calculate them embership degree fuzzy using the following equation : Where : : Fuzzy membership value : the value of adjacent data membership i to the class j : The number of values neighborhoods nearby taken : difference data between the data to the data in the nearest neighbor : weight rank magnitude m > Local Mean K-Nearest Neighbor Local Mean is an approach that works by calculating the vector of test data to the k nearest neighbors of each class. Determination of the target class of the test data using Local Mean approach would be fairer because the distance between the data with each class of targets was calculated. Unlike the vote majority determining the class are only counting the closeness and the majority of the distribution of k nearest class neighborhoods. Gou (2012) states that the editor@iaeme.com
3 Hafizh Al-Kautsar Aidilof, Muhammad Zarlis, Syahril Efendi use of local mean the K-Nearest Neighbor Centroid managed to improve accuracy significantly. Tu (2015) suggested that the main step in Local Mean K-Nearest Neighbor is the smallest Euclidean calculate the test data with each data of each class as much as k nearest neighborhoods. In other words, determination of the class on the K-Nearest Neighbor which use a majority vote to be replaced by considering the proximity characteristic of each class that have been calculated proximity as much as k. Pan (2016) formulate the steps of decisionmaking by local mean the K-Nearest Neighbor as follows : Define training data of ech class Count vector local mean Classify into classes where the Euclidean distance between vectors are local mean and minimum values 3. PROPOSE METHOD Fuzzyfication performed to calculate the degree of membership of the test data of each class of training data. While local mean is used to calculate closeness between the test data to characterize each class nearby. Modeling is done by dividing the value of the membership of Local Mean weighted so that the proximity characteristic fuzzyfikasi result will be more meetings and looks increasingly separate from other target classes. In the high-density data, this approach will further clarify the position of the test data to be placed on a target class that will facilitate classifier and of course the accuracy of the classification increases. Furthermore, the model algorithm is described as follows : Determine the data training and data testing Calculate the distance between data testing on the training data with Euclidean equation Sort distance from the smallest value to the largest value Taking the data as much as the value k of each class of targets Compute Local Mean Vector of data testing to each training class as much as k Calculate the degree of membership of the data to the respective testing classes training as much as k Calculate the Vector Model of each training class as much as k : Class targets of data testing defined by VectorModel with a minimum value of x editor@iaeme.com
4 Model Fuzzy K-Nearest Neighbor with Local Mean for Pattern Recognition 4. RESULT & DISCUSSION In experiments using iris dataset, conducted the training and testing of each class as much as 30 samples and 20 samples of the obtained results on the class accuracy setosa 93.3%, 86.6% in grade versicolor and virginica 100% in class. In other experiments using simple image data with comparison of data training and data testing about 75% : 25% of the overall data obtained a maximum accuracy of 96.43% with the value of k = 1 both in the Local Mean FKNN and primitive FKNN. The change in accuracy is seen when the value of k is enhanced. The comparison between Local Mean FKNN and primitive FKNN can be seen in the following table: Table 1 Comparison between Local Mean FKNN and FKNN No. Value of k Local Mean FKNN FKNN ,43 96, ,30 96, ,30 96, ,15 92, ,15 89,28 Seen in the table that the values k = 1 to k = 3 show the same accuracy results between LMFKNN and primitive FKNN. Both managed to recognize a simple image pattern with a maximum accuracy of 96.43%. However, when the value of k is increased to 4 which means the algorithm takes the neighborhood with a wider range, LMFKNN only decreases accuracy by 0.15%. This is much different from the primitive FKNN which decreased the accuracy of 3.71% and decreased when given k worth 5. While LMFKNN still persisted in 96.15% accuracy. Herein lies the advantage of local mean approach. The local mean does not attract any data that tends to other classes although it is given a high k value which means the algorithm has a wide range of neighborhoods. As for the distribution of data classification can be seen in the following figure: Figure 1 Value of accuration 96,43% editor@iaeme.com
5 Hafizh Al-Kautsar Aidilof, Muhammad Zarlis, Syahril Efendi Figure 2 Value of accuration 96,30% Figure 3 Value of accuration 96,15% Figure 4 Value of accuration 92,59% editor@iaeme.com
6 Model Fuzzy K-Nearest Neighbor with Local Mean for Pattern Recognition Figure 5 Value of accuration 89,28% At the local mean, the target class determination is performed by calculating the euclidian distance between the attribute values of the test data with the average attribute values as much as k neighborhood of each training data class. The test data will be placed on the target class which has the highest mean local vector value. In the k fuzzy k-nearest primitive neighbor, the determination of the target class is performed by calculating the degree of membership of a test pattern against each target class. The target class with the largest degree of degree of membership for a test pattern will be selected as the target class for the test pattern it is placed. This approach still inherits the nature of the vote majority because in the determination of its class is still determined by the number of dominant classes of nearby neighbors although there is a degree of membership that calculates the weight of test data against each target class. Fuzzy k-nearest neighbor algorithm with local mean using both class determination techniques. Having calculated the degree of membership and the local mean vector of the test pattern on the training pattern, the next step is to divide the mean local vector value by degree of membership. This division is done to minimize the vector distance between test patterns with each pattern of training and with the target class so that it will be clearly visible to the class where a pattern will be recognized. 5. CONCLUSIONS It can concluded that Local Mean Vector on K-Nearest Neighbor divided by a fuzzy membership function can minimize Euclidean distance so there would be a tendency of data testing against a target class. With shrinking Euclidean distance, the more it will bring up the similarity of the data to a class and accuracy of the classification results will be higher. In addition, the local mean model of the Fuzzy K-Nearest Neighbor algorithm can also strengthen the algorithm for the high value of k. High fetching k creates an algorithm competing with each other to retrieve data to fit in a class across a wide range of neighborhoods. With the local mean, class determination is based not only on the distance and degree of membership, but also the average attribute value for each target class editor@iaeme.com
7 REFERENCES Hafizh Al-Kautsar Aidilof, Muhammad Zarlis, Syahril Efendi [1] Beyan, C & Ogul, H A Fuzzy K-NN Approach for Cancer Diagnosis with Microarray Gene Expression Data. [2] Bhatia, N. & Vandana., Survey of Nearest Neighbor Techniques. International Journal of Computer Science and Information Security (IJCSIS) : [3] García-Pedrajas, N. & Ortiz-Boyer, D Boosting K-Nearest Neighbor Classifier By Means Of Input Space Projection. Expert System With Application : [4] Gou, J., Yi, Z., Du, L. & Xiong, T A Local Mean-Based k-nearest Centroid Neighbor Classifier. The Computer Journal 55 (9) : [5] Jabbar, M.A., Deekshatulu, B.L. & Chandra. P Classification of Heart Disease Using K- Nearest Neighbor and Genetic Algorithm. International Conference on Computational Intelligence: Modeling Techniques and Applications (CIMTA) : [6] Keller, J.M., Gray, M.R. & Givens, J.A. A Fuzzy K-Nearest Neighbor Algorithm. IEEE Transactions on Systems, Man, and Cybernetics SMC-15 (4) : [7] Ougiaroglou, S. & Evangelidis, G Fast and Accuratek-Nearest Neighbor Classification using Prototype Selection by Clustering. Panhellenic Conference on Informatics. [8] Pan, Z., Wang, Y. & Ku, W A New K-Harmonic Nearest Neighbor Classifier based on the Multi Local Means. [9] Pan, Z., Wang, Y. & Ku, W A New General Nearest Neighbor Classification Based On The Mutual Neighborhood Information. Knowledge-Based Systems : [10] Raikwal, J.S. & Saxena, K Performance Evaluation of SVM and K-Nearest Neighbor Algorithm over Medical Dataset. International Journal of Computer Applications 50 (14) : [11] Rosyid, H., Prasetyo, E. & Agustin, S Perbaikan Akurasi Fuzzy K-Nearest Neighbor In Every Class menggunakan Fungsi Kernel. Seminar Nasional Teknologi Informasi dan Multimedia 2013, pp [12] Sánchez, A.S., Iglesias-Rodríguez, F.J., Fernándes, P.R. & Juez, F.J.de.C Applying The K-Nearest Neighbor Technique To The Classification Of Workers According To Their Risk Of Suffering Musculoskeletal Disorders. International Journal of Indsutrial Ergonomics : 1-8. [13] Song, Y., Liang, J., Lu, J. & Zhao, X An Efficient Instance Selection Algorithm For K Nearest Neighbor Regression. Neurocomputing : 26-34, Volume : 251. [14] Tu, L., Wei, H. & Ai, L Galaxy and Quasar Classificication Based on Local Mean based K-Nearest Neighbor Method. IEEE, /15. [15] Wang. J., Neskovic. P. & Cooper L.N., Improving Nearest Neighbor Rule With A Simple Adaptive Distance Measure. Pattern Recognition Letter : , vol 28. [16] Wu, X. & Kumar, V The Top Ten Algorithms in Data Mining. CRC Press : Boca Raton, USA. [17] Zbancioc, M. & Feraru, S.M Emotion Recognition of the SROL Romanian Database using Fuzzy KNN Algorithm. IEEE, / editor@iaeme.com
Improving the accuracy of k-nearest neighbor using local mean based and distance weight
Journal of Physics: Conference Series PAPER OPEN ACCESS Improving the accuracy of k-nearest neighbor using local mean based and distance weight To cite this article: K U Syaliman et al 2018 J. Phys.: Conf.
More informationUsing a genetic algorithm for editing k-nearest neighbor classifiers
Using a genetic algorithm for editing k-nearest neighbor classifiers R. Gil-Pita 1 and X. Yao 23 1 Teoría de la Señal y Comunicaciones, Universidad de Alcalá, Madrid (SPAIN) 2 Computer Sciences Department,
More informationA STUDY OF SOME DATA MINING CLASSIFICATION TECHNIQUES
A STUDY OF SOME DATA MINING CLASSIFICATION TECHNIQUES Narsaiah Putta Assistant professor Department of CSE, VASAVI College of Engineering, Hyderabad, Telangana, India Abstract Abstract An Classification
More informationNaïve Bayes for text classification
Road Map Basic concepts Decision tree induction Evaluation of classifiers Rule induction Classification using association rules Naïve Bayesian classification Naïve Bayes for text classification Support
More informationIntroduction to Artificial Intelligence
Introduction to Artificial Intelligence COMP307 Machine Learning 2: 3-K Techniques Yi Mei yi.mei@ecs.vuw.ac.nz 1 Outline K-Nearest Neighbour method Classification (Supervised learning) Basic NN (1-NN)
More informationA Fuzzy C-means Clustering Algorithm Based on Pseudo-nearest-neighbor Intervals for Incomplete Data
Journal of Computational Information Systems 11: 6 (2015) 2139 2146 Available at http://www.jofcis.com A Fuzzy C-means Clustering Algorithm Based on Pseudo-nearest-neighbor Intervals for Incomplete Data
More informationThe Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem
Int. J. Advance Soft Compu. Appl, Vol. 9, No. 1, March 2017 ISSN 2074-8523 The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem Loc Tran 1 and Linh Tran
More informationUsing Center Representation and Variance Effect on K-NN Classification
Using Center Representation and Variance Effect on K-NN Classification Tamer TULGAR Department of Computer Engineering Girne American University Girne, T.R.N.C., Mersin 10 TURKEY tamertulgar@gau.edu.tr
More informationAutomatic Classification of Audio Data
Automatic Classification of Audio Data Carlos H. C. Lopes, Jaime D. Valle Jr. & Alessandro L. Koerich IEEE International Conference on Systems, Man and Cybernetics The Hague, The Netherlands October 2004
More informationAn Automatic 3D Face Model Segmentation for Acquiring Weight Motion Area
An Automatic 3D Face Model Segmentation for Acquiring Weight Motion Area Rio Caesar Suyoto Samuel Gandang Gunanto Magister Informatics Engineering Atma Jaya Yogyakarta University Sleman, Indonesia Magister
More informationFEATURE EXTRACTION TECHNIQUES USING SUPPORT VECTOR MACHINES IN DISEASE PREDICTION
FEATURE EXTRACTION TECHNIQUES USING SUPPORT VECTOR MACHINES IN DISEASE PREDICTION Sandeep Kaur 1, Dr. Sheetal Kalra 2 1,2 Computer Science Department, Guru Nanak Dev University RC, Jalandhar(India) ABSTRACT
More informationA Classifier with the Function-based Decision Tree
A Classifier with the Function-based Decision Tree Been-Chian Chien and Jung-Yi Lin Institute of Information Engineering I-Shou University, Kaohsiung 84008, Taiwan, R.O.C E-mail: cbc@isu.edu.tw, m893310m@isu.edu.tw
More informationREMOVAL OF REDUNDANT AND IRRELEVANT DATA FROM TRAINING DATASETS USING SPEEDY FEATURE SELECTION METHOD
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology ISSN 2320 088X IMPACT FACTOR: 5.258 IJCSMC,
More informationFeature weighting using particle swarm optimization for learning vector quantization classifier
Journal of Physics: Conference Series PAPER OPEN ACCESS Feature weighting using particle swarm optimization for learning vector quantization classifier To cite this article: A Dongoran et al 2018 J. Phys.:
More informationQUERY REGION DETERMINATION BASED ON REGION IMPORTANCE INDEX AND RELATIVE POSITION FOR REGION-BASED IMAGE RETRIEVAL
International Journal of Technology (2016) 4: 654-662 ISSN 2086-9614 IJTech 2016 QUERY REGION DETERMINATION BASED ON REGION IMPORTANCE INDEX AND RELATIVE POSITION FOR REGION-BASED IMAGE RETRIEVAL Pasnur
More informationOn Classification: An Empirical Study of Existing Algorithms Based on Two Kaggle Competitions
On Classification: An Empirical Study of Existing Algorithms Based on Two Kaggle Competitions CAMCOS Report Day December 9th, 2015 San Jose State University Project Theme: Classification The Kaggle Competition
More informationBUBBLE ALGORITHM FOR THE REDUCTION OF REFERENCE 1. INTRODUCTION
JOURNAL OF MEDICAL INFORMATICS & TECHNOLOGIES Vol. 16/2010, ISSN 1642-6037 pattern recognition, nearest neighbour rule, reference set condensation, reference set reduction, bubble algorithms Artur SIERSZEŃ
More informationA Maximal Margin Classification Algorithm Based on Data Field
Send Orders for Reprints to reprints@benthamscience.ae 1088 The Open Cybernetics & Systemics Journal, 2015, 9, 1088-1093 Open Access A Maximal Margin Classification Algorithm Based on Data Field Zhao 1,*,
More informationImproving K-NN Internet Traffic Classification Using Clustering and Principle Component Analysis
Bulletin of Electrical Engineering and Informatics ISSN: 2302-9285 Vol. 6, No. 2, June 2017, pp. 159~165, DOI: 10.11591/eei.v6i2.608 159 Improving K-NN Internet Traffic Classification Using Clustering
More informationCS6716 Pattern Recognition
CS6716 Pattern Recognition Prototype Methods Aaron Bobick School of Interactive Computing Administrivia Problem 2b was extended to March 25. Done? PS3 will be out this real soon (tonight) due April 10.
More informationData Mining Classification: Alternative Techniques. Lecture Notes for Chapter 4. Instance-Based Learning. Introduction to Data Mining, 2 nd Edition
Data Mining Classification: Alternative Techniques Lecture Notes for Chapter 4 Instance-Based Learning Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar Instance Based Classifiers
More informationMachine Learning: Algorithms and Applications Mockup Examination
Machine Learning: Algorithms and Applications Mockup Examination 14 May 2012 FIRST NAME STUDENT NUMBER LAST NAME SIGNATURE Instructions for students Write First Name, Last Name, Student Number and Signature
More informationAvailable Online through
Available Online through www.ijptonline.com ISSN: 0975-766X CODEN: IJPTFI Research Article ANALYSIS OF CT LIVER IMAGES FOR TUMOUR DIAGNOSIS BASED ON CLUSTERING TECHNIQUE AND TEXTURE FEATURES M.Krithika
More informationMultiple Classifier Fusion using k-nearest Localized Templates
Multiple Classifier Fusion using k-nearest Localized Templates Jun-Ki Min and Sung-Bae Cho Department of Computer Science, Yonsei University Biometrics Engineering Research Center 134 Shinchon-dong, Sudaemoon-ku,
More informationECG782: Multidimensional Digital Signal Processing
ECG782: Multidimensional Digital Signal Processing Object Recognition http://www.ee.unlv.edu/~b1morris/ecg782/ 2 Outline Knowledge Representation Statistical Pattern Recognition Neural Networks Boosting
More informationApplying Supervised Learning
Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains
More informationA PSO-based Generic Classifier Design and Weka Implementation Study
International Forum on Mechanical, Control and Automation (IFMCA 16) A PSO-based Generic Classifier Design and Weka Implementation Study Hui HU1, a Xiaodong MAO1, b Qin XI1, c 1 School of Economics and
More informationData Cleaning and Prototyping Using K-Means to Enhance Classification Accuracy
Data Cleaning and Prototyping Using K-Means to Enhance Classification Accuracy Lutfi Fanani 1 and Nurizal Dwi Priandani 2 1 Department of Computer Science, Brawijaya University, Malang, Indonesia. 2 Department
More informationPerformance Analysis of Data Mining Classification Techniques
Performance Analysis of Data Mining Classification Techniques Tejas Mehta 1, Dr. Dhaval Kathiriya 2 Ph.D. Student, School of Computer Science, Dr. Babasaheb Ambedkar Open University, Gujarat, India 1 Principal
More informationTumor Detection and classification of Medical MRI UsingAdvance ROIPropANN Algorithm
International Journal of Engineering Research and Advanced Technology (IJERAT) DOI:http://dx.doi.org/10.31695/IJERAT.2018.3273 E-ISSN : 2454-6135 Volume.4, Issue 6 June -2018 Tumor Detection and classification
More informationCluster Analysis. Ying Shen, SSE, Tongji University
Cluster Analysis Ying Shen, SSE, Tongji University Cluster analysis Cluster analysis groups data objects based only on the attributes in the data. The main objective is that The objects within a group
More informationOpen Access Research on the Prediction Model of Material Cost Based on Data Mining
Send Orders for Reprints to reprints@benthamscience.ae 1062 The Open Mechanical Engineering Journal, 2015, 9, 1062-1066 Open Access Research on the Prediction Model of Material Cost Based on Data Mining
More informationSupervised vs unsupervised clustering
Classification Supervised vs unsupervised clustering Cluster analysis: Classes are not known a- priori. Classification: Classes are defined a-priori Sometimes called supervised clustering Extract useful
More informationFuzzy C-means Clustering with Temporal-based Membership Function
Indian Journal of Science and Technology, Vol (S()), DOI:./ijst//viS/, December ISSN (Print) : - ISSN (Online) : - Fuzzy C-means Clustering with Temporal-based Membership Function Aseel Mousa * and Yuhanis
More informationOptimization Model of K-Means Clustering Using Artificial Neural Networks to Handle Class Imbalance Problem
IOP Conference Series: Materials Science and Engineering PAPER OPEN ACCESS Optimization Model of K-Means Clustering Using Artificial Neural Networks to Handle Class Imbalance Problem To cite this article:
More informationKTH ROYAL INSTITUTE OF TECHNOLOGY. Lecture 14 Machine Learning. K-means, knn
KTH ROYAL INSTITUTE OF TECHNOLOGY Lecture 14 Machine Learning. K-means, knn Contents K-means clustering K-Nearest Neighbour Power Systems Analysis An automated learning approach Understanding states in
More informationData Mining: An experimental approach with WEKA on UCI Dataset
Data Mining: An experimental approach with WEKA on UCI Dataset Ajay Kumar Dept. of computer science Shivaji College University of Delhi, India Indranath Chatterjee Dept. of computer science Faculty of
More informationFeature-Guided K-Means Algorithm for Optimal Image Vector Quantizer Design
Journal of Information Hiding and Multimedia Signal Processing c 2017 ISSN 2073-4212 Ubiquitous International Volume 8, Number 6, November 2017 Feature-Guided K-Means Algorithm for Optimal Image Vector
More informationNearest Cluster Classifier
Nearest Cluster Classifier Hamid Parvin, Moslem Mohamadi, Sajad Parvin, Zahra Rezaei, and Behrouz Minaei Nourabad Mamasani Branch, Islamic Azad University, Nourabad Mamasani, Iran hamidparvin@mamasaniiau.ac.ir,
More informationAn Empirical Study on Lazy Multilabel Classification Algorithms
An Empirical Study on Lazy Multilabel Classification Algorithms Eleftherios Spyromitros, Grigorios Tsoumakas and Ioannis Vlahavas Machine Learning & Knowledge Discovery Group Department of Informatics
More informationContent Based Image Retrieval system with a combination of Rough Set and Support Vector Machine
Shahabi Lotfabadi, M., Shiratuddin, M.F. and Wong, K.W. (2013) Content Based Image Retrieval system with a combination of rough set and support vector machine. In: 9th Annual International Joint Conferences
More informationInternational Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X
Analysis about Classification Techniques on Categorical Data in Data Mining Assistant Professor P. Meena Department of Computer Science Adhiyaman Arts and Science College for Women Uthangarai, Krishnagiri,
More informationImplementation of Modified K-Nearest Neighbor for Diagnosis of Liver Patients
Implementation of Modified K-Nearest Neighbor for Diagnosis of Liver Patients Alwis Nazir, Lia Anggraini, Elvianti, Suwanto Sanjaya, Fadhilla Syafria Department of Informatics, Faculty of Science and Technology
More informationUSING OF THE K NEAREST NEIGHBOURS ALGORITHM (k-nns) IN THE DATA CLASSIFICATION
USING OF THE K NEAREST NEIGHBOURS ALGORITHM (k-nns) IN THE DATA CLASSIFICATION Gîlcă Natalia, Roșia de Amaradia Technological High School, Gorj, ROMANIA Gîlcă Gheorghe, Constantin Brîncuși University from
More informationCotton Texture Segmentation Based On Image Texture Analysis Using Gray Level Co-occurrence Matrix (GLCM) And Euclidean Distance
Cotton Texture Segmentation Based On Image Texture Analysis Using Gray Level Co-occurrence Matrix (GLCM) And Euclidean Distance Farell Dwi Aferi 1, Tito Waluyo Purboyo 2 and Randy Erfa Saputra 3 1 College
More informationSemi-Supervised Clustering with Partial Background Information
Semi-Supervised Clustering with Partial Background Information Jing Gao Pang-Ning Tan Haibin Cheng Abstract Incorporating background knowledge into unsupervised clustering algorithms has been the subject
More informationColor-Based Classification of Natural Rock Images Using Classifier Combinations
Color-Based Classification of Natural Rock Images Using Classifier Combinations Leena Lepistö, Iivari Kunttu, and Ari Visa Tampere University of Technology, Institute of Signal Processing, P.O. Box 553,
More informationContents. Preface to the Second Edition
Preface to the Second Edition v 1 Introduction 1 1.1 What Is Data Mining?....................... 4 1.2 Motivating Challenges....................... 5 1.3 The Origins of Data Mining....................
More informationKeywords Binary Linked Object, Binary silhouette, Fingertip Detection, Hand Gesture Recognition, k-nn algorithm.
Volume 7, Issue 5, May 2017 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Hand Gestures Recognition
More informationInstance-Based Representations. k-nearest Neighbor. k-nearest Neighbor. k-nearest Neighbor. exemplars + distance measure. Challenges.
Instance-Based Representations exemplars + distance measure Challenges. algorithm: IB1 classify based on majority class of k nearest neighbors learned structure is not explicitly represented choosing k
More informationK-means clustering based filter feature selection on high dimensional data
International Journal of Advances in Intelligent Informatics ISSN: 2442-6571 Vol 2, No 1, March 2016, pp. 38-45 38 K-means clustering based filter feature selection on high dimensional data Dewi Pramudi
More informationKernels + K-Means Introduction to Machine Learning. Matt Gormley Lecture 29 April 25, 2018
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Kernels + K-Means Matt Gormley Lecture 29 April 25, 2018 1 Reminders Homework 8:
More informationNearest Cluster Classifier
Nearest Cluster Classifier Hamid Parvin, Moslem Mohamadi, Sajad Parvin, Zahra Rezaei, Behrouz Minaei Nourabad Mamasani Branch Islamic Azad University Nourabad Mamasani, Iran hamidparvin@mamasaniiau.ac.ir,
More informationINFORMATION-THEORETIC OUTLIER DETECTION FOR LARGE-SCALE CATEGORICAL DATA
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 4, April 2015,
More informationThe Curse of Dimensionality
The Curse of Dimensionality ACAS 2002 p1/66 Curse of Dimensionality The basic idea of the curse of dimensionality is that high dimensional data is difficult to work with for several reasons: Adding more
More informationThe Comparative Study of Machine Learning Algorithms in Text Data Classification*
The Comparative Study of Machine Learning Algorithms in Text Data Classification* Wang Xin School of Science, Beijing Information Science and Technology University Beijing, China Abstract Classification
More informationCHAPTER 4: CLUSTER ANALYSIS
CHAPTER 4: CLUSTER ANALYSIS WHAT IS CLUSTER ANALYSIS? A cluster is a collection of data-objects similar to one another within the same group & dissimilar to the objects in other groups. Cluster analysis
More informationMachine Learning in Biology
Università degli studi di Padova Machine Learning in Biology Luca Silvestrin (Dottorando, XXIII ciclo) Supervised learning Contents Class-conditional probability density Linear and quadratic discriminant
More information9 Classification: KNN and SVM
CSE4334/5334 Data Mining 9 Classification: KNN and SVM Chengkai Li Department of Computer Science and Engineering University of Texas at Arlington Fall 2017 (Slides courtesy of Pang-Ning Tan, Michael Steinbach
More informationHARD, SOFT AND FUZZY C-MEANS CLUSTERING TECHNIQUES FOR TEXT CLASSIFICATION
HARD, SOFT AND FUZZY C-MEANS CLUSTERING TECHNIQUES FOR TEXT CLASSIFICATION 1 M.S.Rekha, 2 S.G.Nawaz 1 PG SCALOR, CSE, SRI KRISHNADEVARAYA ENGINEERING COLLEGE, GOOTY 2 ASSOCIATE PROFESSOR, SRI KRISHNADEVARAYA
More informationFUZZY KERNEL K-MEDOIDS ALGORITHM FOR MULTICLASS MULTIDIMENSIONAL DATA CLASSIFICATION
FUZZY KERNEL K-MEDOIDS ALGORITHM FOR MULTICLASS MULTIDIMENSIONAL DATA CLASSIFICATION 1 ZUHERMAN RUSTAM, 2 AINI SURI TALITA 1 Senior Lecturer, Department of Mathematics, Faculty of Mathematics and Natural
More informationOverview Citation. ML Introduction. Overview Schedule. ML Intro Dataset. Introduction to Semi-Supervised Learning Review 10/4/2010
INFORMATICS SEMINAR SEPT. 27 & OCT. 4, 2010 Introduction to Semi-Supervised Learning Review 2 Overview Citation X. Zhu and A.B. Goldberg, Introduction to Semi- Supervised Learning, Morgan & Claypool Publishers,
More informationReview of feature selection techniques in bioinformatics by Yvan Saeys, Iñaki Inza and Pedro Larrañaga.
Americo Pereira, Jan Otto Review of feature selection techniques in bioinformatics by Yvan Saeys, Iñaki Inza and Pedro Larrañaga. ABSTRACT In this paper we want to explain what feature selection is and
More informationLEARNING WEIGHTS OF FUZZY RULES BY USING GRAVITATIONAL SEARCH ALGORITHM
International Journal of Innovative Computing, Information and Control ICIC International c 2013 ISSN 1349-4198 Volume 9, Number 4, April 2013 pp. 1593 1601 LEARNING WEIGHTS OF FUZZY RULES BY USING GRAVITATIONAL
More informationComparative analysis of classifier algorithm in data mining Aikjot Kaur Narula#, Dr.Raman Maini*
Comparative analysis of classifier algorithm in data mining Aikjot Kaur Narula#, Dr.Raman Maini* #Student, Department of Computer Engineering, Punjabi university Patiala, India, aikjotnarula@gmail.com
More informationCS 584 Data Mining. Classification 1
CS 584 Data Mining Classification 1 Classification: Definition Given a collection of records (training set ) Each record contains a set of attributes, one of the attributes is the class. Find a model for
More informationEncoding Words into String Vectors for Word Categorization
Int'l Conf. Artificial Intelligence ICAI'16 271 Encoding Words into String Vectors for Word Categorization Taeho Jo Department of Computer and Information Communication Engineering, Hongik University,
More informationAN IMPROVED DENSITY BASED k-means ALGORITHM
AN IMPROVED DENSITY BASED k-means ALGORITHM Kabiru Dalhatu 1 and Alex Tze Hiang Sim 2 1 Department of Computer Science, Faculty of Computing and Mathematical Science, Kano University of Science and Technology
More informationA New Method For Forecasting Enrolments Combining Time-Variant Fuzzy Logical Relationship Groups And K-Means Clustering
A New Method For Forecasting Enrolments Combining Time-Variant Fuzzy Logical Relationship Groups And K-Means Clustering Nghiem Van Tinh 1, Vu Viet Vu 1, Tran Thi Ngoc Linh 1 1 Thai Nguyen University of
More informationCAMCOS Report Day. December 9 th, 2015 San Jose State University Project Theme: Classification
CAMCOS Report Day December 9 th, 2015 San Jose State University Project Theme: Classification On Classification: An Empirical Study of Existing Algorithms based on two Kaggle Competitions Team 1 Team 2
More informationNearby Search Indekos Based Android Using A Star (A*) Algorithm
Journal of Physics: Conference Series PAPER OPEN ACCESS Nearby Search Indekos Based Android Using A Star (A*) Algorithm To cite this article: B Siregar et al 2018 J. Phys.: Conf. Ser. 978 012084 View the
More informationAn Efficient Semantic Image Retrieval based on Color and Texture Features and Data Mining Techniques
An Efficient Semantic Image Retrieval based on Color and Texture Features and Data Mining Techniques Doaa M. Alebiary Department of computer Science, Faculty of computers and informatics Benha University
More informationClustering & Classification (chapter 15)
Clustering & Classification (chapter 5) Kai Goebel Bill Cheetham RPI/GE Global Research goebel@cs.rpi.edu cheetham@cs.rpi.edu Outline k-means Fuzzy c-means Mountain Clustering knn Fuzzy knn Hierarchical
More informationCSE 6242 A / CX 4242 DVA. March 6, Dimension Reduction. Guest Lecturer: Jaegul Choo
CSE 6242 A / CX 4242 DVA March 6, 2014 Dimension Reduction Guest Lecturer: Jaegul Choo Data is Too Big To Analyze! Limited memory size! Data may not be fitted to the memory of your machine! Slow computation!
More informationString Vector based KNN for Text Categorization
458 String Vector based KNN for Text Categorization Taeho Jo Department of Computer and Information Communication Engineering Hongik University Sejong, South Korea tjo018@hongik.ac.kr Abstract This research
More informationMass Classification Method in Mammogram Using Fuzzy K-Nearest Neighbour Equality
Mass Classification Method in Mammogram Using Fuzzy K-Nearest Neighbour Equality Abstract: Mass classification of objects is an important area of research and application in a variety of fields. In this
More informationIntroduction to Support Vector Machines
Introduction to Support Vector Machines CS 536: Machine Learning Littman (Wu, TA) Administration Slides borrowed from Martin Law (from the web). 1 Outline History of support vector machines (SVM) Two classes,
More informationAdaptive Gesture Recognition System Integrating Multiple Inputs
Adaptive Gesture Recognition System Integrating Multiple Inputs Master Thesis - Colloquium Tobias Staron University of Hamburg Faculty of Mathematics, Informatics and Natural Sciences Technical Aspects
More informationk Nearest Neighbors Super simple idea! Instance-based learning as opposed to model-based (no pre-processing)
k Nearest Neighbors k Nearest Neighbors To classify an observation: Look at the labels of some number, say k, of neighboring observations. The observation is then classified based on its nearest neighbors
More informationS. Sreenivasan Research Scholar, School of Advanced Sciences, VIT University, Chennai Campus, Vandalur-Kelambakkam Road, Chennai, Tamil Nadu, India
International Journal of Civil Engineering and Technology (IJCIET) Volume 9, Issue 10, October 2018, pp. 1322 1330, Article ID: IJCIET_09_10_132 Available online at http://www.iaeme.com/ijciet/issues.asp?jtype=ijciet&vtype=9&itype=10
More informationKeyword Extraction by KNN considering Similarity among Features
64 Int'l Conf. on Advances in Big Data Analytics ABDA'15 Keyword Extraction by KNN considering Similarity among Features Taeho Jo Department of Computer and Information Engineering, Inha University, Incheon,
More informationA REVIEW ON VARIOUS APPROACHES OF CLUSTERING IN DATA MINING
A REVIEW ON VARIOUS APPROACHES OF CLUSTERING IN DATA MINING Abhinav Kathuria Email - abhinav.kathuria90@gmail.com Abstract: Data mining is the process of the extraction of the hidden pattern from the data
More informationK-modes Clustering Algorithm for Categorical Data
K-modes Clustering Algorithm for Categorical Data Neha Sharma Samrat Ashok Technological Institute Department of Information Technology, Vidisha, India Nirmal Gaud Samrat Ashok Technological Institute
More informationVECTOR SPACE CLASSIFICATION
VECTOR SPACE CLASSIFICATION Christopher D. Manning, Prabhakar Raghavan and Hinrich Schütze, Introduction to Information Retrieval, Cambridge University Press. Chapter 14 Wei Wei wwei@idi.ntnu.no Lecture
More informationK Nearest Neighbor Wrap Up K- Means Clustering. Slides adapted from Prof. Carpuat
K Nearest Neighbor Wrap Up K- Means Clustering Slides adapted from Prof. Carpuat K Nearest Neighbor classification Classification is based on Test instance with Training Data K: number of neighbors that
More informationUnsupervised Learning : Clustering
Unsupervised Learning : Clustering Things to be Addressed Traditional Learning Models. Cluster Analysis K-means Clustering Algorithm Drawbacks of traditional clustering algorithms. Clustering as a complex
More informationOCR For Handwritten Marathi Script
International Journal of Scientific & Engineering Research Volume 3, Issue 8, August-2012 1 OCR For Handwritten Marathi Script Mrs.Vinaya. S. Tapkir 1, Mrs.Sushma.D.Shelke 2 1 Maharashtra Academy Of Engineering,
More informationUnsupervised Learning
Unsupervised Learning Unsupervised learning Until now, we have assumed our training samples are labeled by their category membership. Methods that use labeled samples are said to be supervised. However,
More informationNearest Neighbor Classifiers
Nearest Neighbor Classifiers TNM033 Data Mining Techniques Linköping University 2009-12-04 When I see a bird that walks like a duck and swims like a duck and quacks like a duck, I call that bird a duck.
More informationAUTOMATIC CLUSTERING AND OPTIMIZED FUZZY LOGICAL RELATIONSHIPS FOR MINIMUM LIVING NEEDS FORECASTING
Journal of Environmental Engineering & Sustainable Technology JEEST http://jeest.ub.ac.id AUTOMATIC CLUSTERING AND OPTIMIZED FUZZY LOGICAL RELATIONSHIPS FOR MINIMUM LIVING NEEDS FORECASTING Yusuf Priyo
More informationAn Unsupervised Technique for Statistical Data Analysis Using Data Mining
International Journal of Information Sciences and Application. ISSN 0974-2255 Volume 5, Number 1 (2013), pp. 11-20 International Research Publication House http://www.irphouse.com An Unsupervised Technique
More informationAvailable online at ScienceDirect. Procedia Computer Science 35 (2014 )
Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 35 (2014 ) 388 396 18 th International Conference on Knowledge-Based and Intelligent Information & Engineering Systems
More informationTopic 1 Classification Alternatives
Topic 1 Classification Alternatives [Jiawei Han, Micheline Kamber, Jian Pei. 2011. Data Mining Concepts and Techniques. 3 rd Ed. Morgan Kaufmann. ISBN: 9380931913.] 1 Contents 2. Classification Using Frequent
More informationA CRITIQUE ON IMAGE SEGMENTATION USING K-MEANS CLUSTERING ALGORITHM
A CRITIQUE ON IMAGE SEGMENTATION USING K-MEANS CLUSTERING ALGORITHM S.Jaipriya, Assistant professor, Department of ECE, Sri Krishna College of Technology R.Abimanyu, UG scholars, Department of ECE, Sri
More informationDistribution-free Predictive Approaches
Distribution-free Predictive Approaches The methods discussed in the previous sections are essentially model-based. Model-free approaches such as tree-based classification also exist and are popular for
More informationA density-based approach for instance selection
2015 IEEE 27th International Conference on Tools with Artificial Intelligence A density-based approach for instance selection Joel Luis Carbonera Institute of Informatics Universidade Federal do Rio Grande
More informationCHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION
CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION 6.1 INTRODUCTION Fuzzy logic based computational techniques are becoming increasingly important in the medical image analysis arena. The significant
More informationAccelerating Unique Strategy for Centroid Priming in K-Means Clustering
IJIRST International Journal for Innovative Research in Science & Technology Volume 3 Issue 07 December 2016 ISSN (online): 2349-6010 Accelerating Unique Strategy for Centroid Priming in K-Means Clustering
More informationClassification: Decision Trees
Classification: Decision Trees IST557 Data Mining: Techniques and Applications Jessie Li, Penn State University 1 Decision Tree Example Will a pa)ent have high-risk based on the ini)al 24-hour observa)on?
More informationA Study on K-Means Clustering in Text Mining Using Python
International Journal of Computer Systems (ISSN: 2394-1065), Volume 03 Issue 08, August, 2016 Available at http://www.ijcsonline.com/ Dr. (Ms). Ananthi Sheshasayee 1, Ms. G. Thailambal 2 1 Head and Associate
More information