Text clustering based on a divide and merge strategy

Similar documents
Clustering. Chapter 10 in Introduction to statistical learning

COMPARISON OF DENSITY-BASED CLUSTERING ALGORITHMS

Hierarchical Document Clustering

A Weighted Majority Voting based on Normalized Mutual Information for Cluster Analysis

Available online at ScienceDirect. Procedia Computer Science 89 (2016 )

Clustering Algorithms for Data Stream

Unsupervised learning on Color Images

Available online at ScienceDirect. Procedia Computer Science 89 (2016 )

DENSITY BASED AND PARTITION BASED CLUSTERING OF UNCERTAIN DATA BASED ON KL-DIVERGENCE SIMILARITY MEASURE

An Efficient Hash-based Association Rule Mining Approach for Document Clustering

Outlier Detection Using Unsupervised and Semi-Supervised Technique on High Dimensional Data

A Naïve Soft Computing based Approach for Gene Expression Data Analysis

Enhancing K-means Clustering Algorithm with Improved Initial Center

Discovering Advertisement Links by Using URL Text

Document Clustering: Comparison of Similarity Measures

Chinese text clustering algorithm based k-means

Clustering in Data Mining

An Improved KNN Classification Algorithm based on Sampling

Normalization based K means Clustering Algorithm

Data Clustering Hierarchical Clustering, Density based clustering Grid based clustering

Keywords Clustering, Goals of clustering, clustering techniques, clustering algorithms.

Video annotation based on adaptive annular spatial partition scheme

Clustering Documents in Large Text Corpora

CHAPTER 3 ASSOCIATON RULE BASED CLUSTERING

Yunfeng Zhang 1, Huan Wang 2, Jie Zhu 1 1 Computer Science & Engineering Department, North China Institute of Aerospace

Index Terms:- Document classification, document clustering, similarity measure, accuracy, classifiers, clustering algorithms.

Available online at ScienceDirect. Procedia Engineering 99 (2015 )

Available online at ScienceDirect. Procedia Engineering 161 (2016 ) Bohdan Stawiski a, *, Tomasz Kania a

Available online at ScienceDirect. Procedia Computer Science 89 (2016 )

Heterogeneous Density Based Spatial Clustering of Application with Noise

An Improvement of Centroid-Based Classification Algorithm for Text Classification

An Efficient QBIR system using Adaptive segmentation and multiple features

ScienceDirect. An Efficient Association Rule Based Clustering of XML Documents

AN IMPROVED DENSITY BASED k-means ALGORITHM

K-Means Clustering With Initial Centroids Based On Difference Operator

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

Clustering Part 4 DBSCAN

C-NBC: Neighborhood-Based Clustering with Constraints

Information Retrieval using Pattern Deploying and Pattern Evolving Method for Text Mining

A Hierarchical Document Clustering Approach with Frequent Itemsets

Keyword Extraction by KNN considering Similarity among Features

Clustering part II 1

Improving Suffix Tree Clustering Algorithm for Web Documents

CS229 Final Project: k-means Algorithm

Unsupervised Learning and Clustering

Implementation of Parallel CASINO Algorithm Based on MapReduce. Li Zhang a, Yijie Shi b

Available online at ScienceDirect. Procedia Computer Science 59 (2015 )

Analysis and Extensions of Popular Clustering Algorithms

Available online at ScienceDirect. Procedia Computer Science 45 (2015 )

Clustering Algorithms In Data Mining

Text Documents clustering using K Means Algorithm

HOT asax: A Novel Adaptive Symbolic Representation for Time Series Discords Discovery

Balanced COD-CLARANS: A Constrained Clustering Algorithm to Optimize Logistics Distribution Network

DS504/CS586: Big Data Analytics Big Data Clustering II

Clustering: An art of grouping related objects

Unsupervised Learning and Clustering

ScienceDirect. Enhanced Associative Classification of XML Documents Supported by Semantic Concepts

University of Florida CISE department Gator Engineering. Clustering Part 4

CHAPTER 5 OPTIMAL CLUSTER-BASED RETRIEVAL

Semi supervised clustering for Text Clustering

String Vector based KNN for Text Categorization

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

Clustering. Informal goal. General types of clustering. Applications: Clustering in information search and analysis. Example applications in search

A Review on Cluster Based Approach in Data Mining

CHAPTER 4: CLUSTER ANALYSIS

CHAPTER 7. PAPER 3: EFFICIENT HIERARCHICAL CLUSTERING OF LARGE DATA SETS USING P-TREES

International Journal of Computer Engineering and Applications, Volume VIII, Issue III, Part I, December 14

To Enhance Projection Scalability of Item Transactions by Parallel and Partition Projection using Dynamic Data Set

DS504/CS586: Big Data Analytics Big Data Clustering II

Data Clustering With Leaders and Subleaders Algorithm

Recommendation System Using Yelp Data CS 229 Machine Learning Jia Le Xu, Yingran Xu

Theme Identification in RDF Graphs

The Effect of Word Sampling on Document Clustering


Empirical Analysis of Data Clustering Algorithms

CSE 5243 INTRO. TO DATA MINING

Unsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi

Web Page Classification using FP Growth Algorithm Akansha Garg,Computer Science Department Swami Vivekanad Subharti University,Meerut, India

Dynamic Clustering of Data with Modified K-Means Algorithm

Classifying Documents by Distributed P2P Clustering

Available online at ScienceDirect. Procedia Computer Science 35 (2014 )

[Gidhane* et al., 5(7): July, 2016] ISSN: IC Value: 3.00 Impact Factor: 4.116

Detecting Clusters and Outliers for Multidimensional

Mining Quantitative Association Rules on Overlapped Intervals

Available online at ScienceDirect. Procedia Technology 17 (2014 )

Available online at ScienceDirect. Procedia Computer Science 60 (2015 )

An Efficient Clustering for Crime Analysis

Clustering Large Dynamic Datasets Using Exemplar Points

Density Based Clustering using Modified PSO based Neighbor Selection

Using Association Rules for Better Treatment of Missing Values

Efficiently Mining Positive Correlation Rules

Cluster Analysis. Ying Shen, SSE, Tongji University

K-means algorithm and its application for clustering companies listed in Zhejiang province

CHAPTER 3 A FAST K-MODES CLUSTERING ALGORITHM TO WAREHOUSE VERY LARGE HETEROGENEOUS MEDICAL DATABASES

News Filtering and Summarization System Architecture for Recognition and Summarization of News Pages

Customer Clustering using RFM analysis

CIS UDEL Working Notes on ImageCLEF 2015: Compound figure detection task

Available online at ScienceDirect. Procedia Computer Science 96 (2016 )

Domain Specific Search Engine for Students

Total cost of ownership for application replatform by open-source SW

Transcription:

Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 55 (2015 ) 825 832 Information Technology and Quantitative Management (ITQM 2015) Text clustering based on a divide and merge strategy Man Yuan a,b, Yong Shi b,c, * a China Huarong Asset Management CO., LTD., Beijing, China b Research Center on Fictitious Economy & Data Science, Chinese Academy of Sciences, Univiersity of Chinese Academy of Sciences, Beijing, China c Key Lab of Big Data Mining and Knowledge Management, Chinese Academy of Sciences, Beijing, 100190, China. Abstract A text clustering algorithm is proposed to overcome the drawback of division based clustering method on sensitivity of estimated class number. Complex features including synonym and co-occurring words are extracted to make a feature space containing more semantic information. Then the divide and merge strategy helps the iteration converge to a reasonable cluster number. Experimental results showed that the dynamically updated center number prevent the deterioration of clustering result when k deviates from the real class numbers. When k is too small or large, the difference of clustering results between FC-DM and k-means is more obvious and FC-DM also outperformed other benchmark algorithms. 2015 Published The Authors. by Elsevier Published B.V. This by Elsevier is an open B.V. access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). Selection and/or peer-review under responsibility of the organizers of ITQM 2015 Peer-review under responsibility of the Organizing Committee of ITQM 2015 Keywords: Text clustering, feature extension, k-means 1. Introduction Text clustering which aims to find un-predefined categories is an important unsupervised method in machine learning. The characteristic of text clustering technology to discover potential concept and topics in unstructured text content makes it compatible for many text management and analysis fields such as summary extraction, semantic analysis and search engine optimization. Division based methods such as k-means [1] and its extensions are most widely used algorithms in text clustering. In these algorithms, the data set is firstly divided into several clusters and each cluster includes at least one document. Based on a certain strategy, a series of iteration update the clusters until all of them satisfy a stable condition. For example, k-means chooses k documents as the original centers of clusters and iterates to update the centers until the criteria function reaches optimal. Each cluster is represented by the centroid or the point closest to the centroid. One of the advantages of division based clustering is the low time complexity and * Corresponding author. E-mail address: yuanman@chamc.com.cn. 1877-0509 2015 Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). Peer-review under responsibility of the Organizing Committee of ITQM 2015 doi:10.1016/j.procs.2015.07.153

826 Man Yuan and Yong Shi / Procedia Computer Science 55 ( 2015 ) 825 832 this makes it suitable for large dataset. If the data set include n documents and k-means iterates l times, the time complexity is O(nkl). However, k-means also suffers drawbacks such as the number k must be estimated in a reasonable scale and the clustering results are greatly affected by the original centers. To solve these problem, Buckshot [2], iterative refinement [3] and k-means++ [4] optimize the initialization of centers. [5] proposed kernel function based k-means to process linear non-separable data. Model based k-means [6] are a combination of model based clustering and k-means, which makes assumption on the data distribution and tries to avoid the noise data by optimizing the fitness of data and the model. Besides, k-means uses Euclidean distance to compute the document similarity, Spherical k-means [7] was proposed using the cosine distance. Bisecting k-means is another popular algorithm in recent years. [8] compared bisecting k-means and k-means and found that bisecting k-means was superior to k-means because k-means usually result in uneven clusters. This paper presents a novel text clustering method based on extended text features and a divide and merge strategy. Firstly, complex features including synonym and co-occurring words are extracted to make a feature space containing more semantic information. Then, the cluster center changes dynamically and either divides or merges by certain rules. In each iteration, the center cluster and the similarity list are recorded to determine whether to undertake the divide or merge step until no more new centers are generated. Finally, experimental evaluation on real text data is conducted to prove the effectiveness of the proposed method. 2. Text clustering based on feature center divide and merge strategy (FC-DM) The notion of feature center divide and merge strategy (FC-DM) is to extract complex features to initialize the centers with more semantic information and in each iteration dynamically divide or merge the centers. The main procedures include:(1) preprocess the text content and extract synonym and co-occurring words.(2) initialize the centers by randomly choose a document with extended features and then select the other k-1 centers with maximum distance.(3) update the cluster centers with the divide and merge strategy until all categories reach stable status. 2.1. Complex feature extraction After preprocessing and indexing, each word can obtain its synonym and co-occurring words as extended features. Consequently, each word can be extended into a rich structure which covers more semantic space. On the first part, synonym means different descriptions on the same topic or semantic meaning. Synonym is very common in human language because people usually choose quite personalized ways to express based on the context, habit or rhetoric in different time and place. The statistic in linguistics shows that the probability that people use the same words for the same meaning is less than 20%. Synonyms bothers text clustering because two different words which has the same meaning usually have differentiation effect. When simply evaluate the similarity by original features, the contribution of synonyms are ignored. So it is necessary to extract synonyms to enhance the original feature space for similarity computation. In this paper, synonym extraction are based on synonym dictionary [9] and verbs, adjective and function are neglected. On the other part, co-occurring words are virtually frequent word set with a support count larger than the threshold value and the extraction is conducted using Apriori algorithm. Considering that most of phases or combinations of words in Chinese consist of 2 words, although there are term sets with more items, under the restraint of above metrics, the numbers are too rare to make substantial influences. So, co-occurring that we extract are double term set from full text content. Besides, if word pair (A,B) has a high support but the count of A is much larger than B, there is still no confidence for the connection between A and B. So, the extension using co-occurring words is virtually a conduction as A B and the confidence threshold min_conf should be considered. Finally, the extraction result of co-occurring words is a confident rule list.

Man Yuan and Yong Shi / Procedia Computer Science 55 ( 2015 ) 825 832 827 As is shown in Figure 1, no matter for the document content or cluster center, the original feature consist a word set in which each word correspond to a rich word structure. The feature set for cluster center is extended while initialization and each document was extended when added into the cluster. The update of cluster center is determined by new documents and will not be extended more. 2.2. Initialization and update of cluster centers Fig. 1. Extended features using synonym and co-occurring words Document centers are initialized with a maximum distance strategy using the Euclidean distance between document and center to make sure that the new added center has low similarities with the existing ones. Assume that the data set contains n documents to classify into k categories, firstly, one of the documents is selected randomly and extended with synonym and co-ocurring words as the center to start up. Then, choose another document with lowest similarity in the remaining n-1 documents, extend its features and add it into the cluster center set. The iteration continues as follows: Algorithm: Cluster Initialization Input: Document set D, number of classes k Output: Center Set 1. Randomly choose a document d, set the extended features as Center 1, add Center 1 in to Center Set and delete it from D. 2. Compute the average distance between each remaining document and all centers in Center Set, set the document with maximum distance as the candidate. 3. Extend the candidate center with synonym and co-occurring words and add it into Center Set as Center i. 4. If the number of centers in Center Set is less than k, iterate step 3. 5. Return Center Set The update of centers follows a k-means strategy which relocate the center by assigning each document from data set. As each iteration ends, according to the distance between document and center, each center reserves an ordered class member list and each document reserves an ordered candidate class list. We can also obtain the following metrics of center C i : minimum distance (Min_Dis): the minimum distance between all the class members and the current center; average distance (Ave_Dis): the average distance between all the class members and the current center; maximum distance (Max_Dis): the maximum distance between all the class members and the current center. Furthermore, taking Ave_Dis as a boundary, core members are defined with a distance ranging in [Min_Dis, Ave_Dis] and frontier members are defined with a distance ranging in [Min_Dis, Ave_Dis].

828 Man Yuan and Yong Shi / Procedia Computer Science 55 ( 2015 ) 825 832 2.3. Divide and merge the centers Each iteration of FC-DM generates a series of centers. Because the precision of clustering is greatly affected by the center initialization and the estimated class number may differ from the real one, it is necessary to divide or merge the class which may consist to many members. One of the key problems here is to evaluate the clustering quality because the information is quite limited without class labels or extra knowledge. In this paper, the clustering quality is judged by the number of core and frontier members mentioned above, and each iteration undertakes dividing or merging under a k-means strategy. Dividing Rules:After each iteration, select the class with maximum documents and determine whether to divide it under the following rules: 1. Select the cluster with maximum document number C max as candidate; 2. According to the center distance list of sample document in C max, choose the Candidate_Center which occurs the most times as the second closest center; 3. From center set excluding C max, choose the center as New_Center which is closest with Candidate_Center 4. Compare the distance between sample document and C max with the distance between sample document and New_Center, if the number of documents which is closer to New_Center is more than C max, divide the class as step 5, or else end this iteration. 5. According to the distances of sample document with New_Center and C max, assign all the documents randomly into the new centers, update all the centers and add New_Center into the center set. 6. Compare the document numbers of New_Centerand C max, choose the larger one as candidate and repeat step 2-5. Merging Rules: It is assumed that if there are too many frontier members in a class and these frontier members are disperse too much to generate a new class, this class should be merged into another one. This procedure is as follows: 1. Choose the class with minimum document number C min as candidate; 2. Get the number of core members in C min as Count_Core and the numer of frontier members as Count_Outer; 3. Get the sample divergence metric of C min ( Count _ Outer - Count _ Core) Count _ Core and the average sample dispersion metrics of other classes Ave( ') 4. if Ave( '), document samples in C min is assumed to have a center divergence and should be merged as step 5. Or else, end the step. 5. Assign all documents in C min randomly into other classes in center set and update all centers. 6. Delete C min in center set. 3. Experiment 3.1. Dataset Sougou Dataset is a collection of news articles provided by Sougou Lab, including 18 categories such as International, Sports, Society, Entertainment etc.. Each article document involve page URL, page ID, news title and page con-tent. The html tags have been removed in the page content. In this paper, we selected 10 classes and each class contains 300 documents. All documents were pre-processed by word segmentation using ICTCLAS [10]. The experiment firstly examined the detailed cluster results. Because the divide and merge strategy dynamically updated the clustering numbers, the divide and merge numbers during iteration as well as the cluster results were discussed. And clustering results under different estimated k were compared. Then FC-DM

Man Yuan and Yong Shi / Procedia Computer Science 55 ( 2015 ) 825 832 829 is compared with other division based clustering methods including k-means and Spherical k-means. The comparison contains clustering precision and stability by running on a random start. Finally, FC-DM was compared with other clustering method to prove its effectiveness. The metric used is F-Measure. 3.2. Experiment Results Based on a divide and merge strategy, FC-DM dynamically update the cluster centers, not only relocate the centers, but also change the cluster numbers by divide or merge the existing classes. One of the advantages of such strategy lies in that the estimated number k usually has great impact on the clustering results, given no background knowledge on data set. Therefore, if the class number can also be updated, the clustering result may get revised when k deviates too much from the real class number. Table 1 shows the entire clustering procedures of FC-DM and we can see that the cluster number may increase or decrease as the iteration carries on. The clustering reached convergence after 7 iterations and the final cluster number is 10. Table 1. Cluster centers in iterations and clustering results of F-Measure(%) while k=10 Iteration Times Cluster Number Precision Recall F-Measure Dividing count Merging count 1 12 64.0 59.6 61.7 3 1 2 13 72.1 67.6 69.8 1 0 3 13 65.5 58.3 61.7 1 1 4 12 73.8 66.2 69.8 0 1 5 11 79.5 72.6 75.9 0 1 6 10 75.9 72.2 74.0 0 1 7 10 79.1 73.5 76.2 0 0 It can be observed from table 1 that the final cluster number may be not the same as the estimated number k. Given different estimated number k, table 2 compared the final cluster number and results after the same iteration times. The result shows that when k is less than the real class number, the revision effect of FC-DM is more obvious. The best clustering result was obtained as F-Measure=79.2% when k=9 and the cluster number is 10. Table 2. Clustering results of F-Measure(%) on different k at the 7 th iteration 3.3. Clustering Comparison k Cluster number Precision Recall F-Measure 8 12 71.5 68.2 69.8 9 10 80.1 78.3 79.2 10 11 79.1 73.5 76.2 11 13 62.9 60.5 61.7 12 11 76.6 72.1 74.3 13 13 64.4 57.6 60.8 The clustering result of division based algorithms such as k-means and its extensions is not unique. Although FC-DM uses the maximum distance principal to initialize a dispersed center set, the procedure still selects document randomly and updates the cluster number, making the cluster results unstable. So, the stability of clustering results was also compared with k-means and Spherical k-means. Table 3 records the clustering results of three algorithms including the cluster number of FC-DM after 8 iterations and the number k is set to be 10. Figure 2 is the comparison of the above data from which we can see FC-DM obtained more stable results than k-means and Spherical k-means. One of the main unstable essence of FC-DM is the dynamic cluster number. The discrepancy of F-Measure between cluster number 12 and 10 is 3%.

830 Man Yuan and Yong Shi / Procedia Computer Science 55 ( 2015 ) 825 832 Table 3. F-Measure (%)Clustering stability of F-Measure(%) while k =10 Iteration Times k-means Spherical k-means FC-DM Cluster Number of FC-DM 1 69 70.2 76.2 11 2 66.2 70.6 74.8 11 3 67.3 71.8 75.8 11 4 66.9 70.9 77.9 10 5 65.7 72.6 75.8 10 6 67.5 68.6 73.4 12 7 67.2 70.2 76.6 10 8 68.7 72.6 75.8 11 Fig. 2. Comparison of clustering results while k =10 Table 4 is the average clustering results of three algorithms given different number k. Figure 3 is the comparison of above data from which we can see FC-DM outperforms the other two and when k=9, the F- Measure of FC-DM is 6.2% greater than Spherical k-means and 10.8 greater than k-means when k=10. Note that by dynamically updating the cluster number, when k deviates from the real class number, FC-DM has a revision effect. In Figure 3, when k increases or decreases, F-Measure of k-means and Spherical k-means drops more than FC-DM. The sensitivity to estimated k of FC-DM is comparatively low. When k=8, F-Measure of FC-DM is 17.5% greater than Spherical k-means and 16.6% greater than k-means. When k=13, the discrepancies are respectively 17.2% and 22.8%. It can be concluded that when k deviates from the optimal class number, FC-DM can effectively prevent the deterioration of clustering results. Table 4. Clustering Results of F-Measure(%) on different k k k-means Spherical k-means FC-DM 8 53.2 52.3 69.8 9 61.8 66.7 78.1 10 67.3 71.9 76.3 11 59.1 63.6 65.7 12 49.7 55.3 74.3 13 37.6 43.6 60.8

Man Yuan and Yong Shi / Procedia Computer Science 55 ( 2015 ) 825 832 831 Fig. 3. Comparison of Clustering Results of F-Measure(%) on different k Table 5 Comparison of Clustering Results of F-Measure(%) with other clustering algorithms k-means Bisecting k-means DBScan UPGMA FC-DM 67.3 72.5 73.5 70.6 78.1 Furthermore, we compared the clustering results of FC-DM and other clustering algorithms including Bisecting k-means, DBScan [11] and UPGMA [12]. DBScan is a popular algorithm based on data density and UPGMA is based on hierarchy. As is shown in table 5, FC-DM outperforms other benchmark algorithms and F-Measure is improved as 4.6%~10.8%. 4. Conclusion This paper presents a text clustering algorithm based on feature extension as well as divide and merge strategy. The original features of document are extended with synonym and co-occurring words to make the sample documents and centers include more semantic information. The divide and merge strategy helps the iteration converges to a reasonable cluster number. Compared with other division based methods, the dynamically updated center number prevents the deterioration of clustering result when k deviates from the real class numbers. The experiment results show that when k is smallest and largest, the difference of clustering results between FC-DM and k-means is more obvious and FC-DM also outperformed other benchmark algorithms. References [1] Huang Z. Extensions to the k-means algorithm for clustering large data sets with categorical values. Data Mining and Knowledge Discovery 1998; 2(3): 283-304. [2] Bellot P, El-Bèze M. Clustering by means of Unsupervised Decision Trees or Hierarchical and K-means-like Algorithm. Proc. of RIAO 2000; p. 344-363. [3] Bradley P S, Fayyad U M. Refining initial points for k-means clustering. Proceedings of the fifteenth international conference on machine learning 1998; p. 66. [4] Arthur D, Vassilvitskii S. k-means++: The advantages of careful seeding. Proceedings of the eighteenth annual ACM-SIAM symposium on Discrete algorithms. Society for Industrial and Applied Mathematics 2007; p. 1027-1035. [5] Schölkopf B, Weston J, Eskin E, et al. A kernel approach for learning from almost orthogonal patterns. Principles of Data Mining and Knowledge Discovery 2002; p. 494-511. [6] Zhong S, Ghosh J. Scalable, balanced model-based clustering. Proc. 3rd SIAM Int. Conf. Data Mining 2003; p. 71-82. [7] Zhong S. Efficient online spherical k-means clustering. IJCNN'05 2005; p. 3180-3185.

832 Man Yuan and Yong Shi / Procedia Computer Science 55 ( 2015 ) 825 832 [8] Steinbach M, Karypis G, Kumar V. A comparison of document clustering techniques. KDD workshop on text mining 2000; 400(1): 525-526. [9] http://ir.hit.edu.cn/demo/ltp/sharing_plan.htm. [10] Zhang, H.P., et al. HHMM-based Chinese lexical analyzer ICTCLAS. Proceedings of the second SIGHAN workshop on Chinese language processing 2003; p. 184-187. [11] Ester M, Kriegel H P, Sander J, et al. A density-based algorithm for discovering clusters in large spatial databases with noise. KDD 1996; p. 226-231. [12] Jain A K, Murty M N, Flynn P J. Data clustering: a review. ACM computing surveys (CSUR) 1999;31(3): 264-323.