An Appropriate Similarity Measure for K-Means Algorithm in Clustering Web Documents S Jaiganesh 1 Dr. P. Jaganathan 2

Similar documents
Document Clustering: Comparison of Similarity Measures

CHAPTER 3 ASSOCIATON RULE BASED CLUSTERING

APPLICATION OF MULTIPLE RANDOM CENTROID (MRC) BASED K-MEANS CLUSTERING ALGORITHM IN INSURANCE A REVIEW ARTICLE

Impact of Term Weighting Schemes on Document Clustering A Review

University of Florida CISE department Gator Engineering. Clustering Part 5

Gene Clustering & Classification

International Journal of Advanced Research in Computer Science and Software Engineering

Concept-Based Document Similarity Based on Suffix Tree Document

Text Data Pre-processing and Dimensionality Reduction Techniques for Document Clustering

A PSO-based Generic Classifier Design and Weka Implementation Study

Text Documents clustering using K Means Algorithm

Criterion Functions for Document Clustering Experiments and Analysis

An Improvement of Centroid-Based Classification Algorithm for Text Classification

Particle Swarm Optimization applied to Pattern Recognition

Cluster Analysis. Angela Montanari and Laura Anderlucci

Hierarchical Clustering Algorithms for Document Datasets

CHAPTER 5 OPTIMAL CLUSTER-BASED RETRIEVAL

An Unsupervised Technique for Statistical Data Analysis Using Data Mining

Enhancing Clustering Results In Hierarchical Approach By Mvs Measures

Nearest Clustering Algorithm for Satellite Image Classification in Remote Sensing Applications

Cluster Analysis. Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX April 2008 April 2010

Dynamic Clustering of Data with Modified K-Means Algorithm

An Improved Document Clustering Approach Using Weighted K-Means Algorithm

ECLT 5810 Clustering

Citation for published version (APA): He, J. (2011). Exploring topic structure: Coherence, diversity and relatedness

A Modified Hierarchical Clustering Algorithm for Document Clustering

ECLT 5810 Clustering

Improving Results and Performance of Collaborative Filtering-based Recommender Systems using Cuckoo Optimization Algorithm

A Study on K-Means Clustering in Text Mining Using Python

A Review of K-mean Algorithm

Understanding Clustering Supervising the unsupervised

Semi-Supervised Clustering with Partial Background Information

SYDE Winter 2011 Introduction to Pattern Recognition. Clustering

Analyzing Outlier Detection Techniques with Hybrid Method

A Naïve Soft Computing based Approach for Gene Expression Data Analysis

Document Clustering using Feature Selection Based on Multiviewpoint and Link Similarity Measure

Enhancing K-means Clustering Algorithm with Improved Initial Center

International Journal of Advanced Research in Computer Science and Software Engineering

NORMALIZATION INDEXING BASED ENHANCED GROUPING K-MEAN ALGORITHM

A Recommender System Based on Improvised K- Means Clustering Algorithm

Keywords: clustering algorithms, unsupervised learning, cluster validity

A Comparative study of Clustering Algorithms using MapReduce in Hadoop

CHAPTER-6 WEB USAGE MINING USING CLUSTERING

[Gidhane* et al., 5(7): July, 2016] ISSN: IC Value: 3.00 Impact Factor: 4.116

Index Terms:- Document classification, document clustering, similarity measure, accuracy, classifiers, clustering algorithms.

Clustering of Data with Mixed Attributes based on Unified Similarity Metric

Space and Cosine Similarity measures for Text Document Clustering Venkata Gopala Rao S. Bhanu Prasad A.

FEATURE EXTRACTION TECHNIQUES USING SUPPORT VECTOR MACHINES IN DISEASE PREDICTION

CSE 5243 INTRO. TO DATA MINING

A Review: Content Base Image Mining Technique for Image Retrieval Using Hybrid Clustering

Unsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi

I. INTRODUCTION II. RELATED WORK.

Collaborative Filtering using Euclidean Distance in Recommendation Engine

COLOR image segmentation is a method of assigning

A HYBRID APPROACH FOR DATA CLUSTERING USING DATA MINING TECHNIQUES

Using Coherence-based Measures to Predict Query Difficulty

Semi supervised clustering for Text Clustering

Courtesy of Prof. Shixia University

Road map. Basic concepts

Clustering and Visualisation of Data

TOWARDS NEW ESTIMATING INCREMENTAL DIMENSIONAL ALGORITHM (EIDA)

An Efficient Analysis for High Dimensional Dataset Using K-Means Hybridization with Ant Colony Optimization Algorithm

Getting to Know Your Data

Available Online through

Clustering Web Documents using Hierarchical Method for Efficient Cluster Formation

Chapter 2 Basic Structure of High-Dimensional Spaces

An Efficient Clustering Method for k-anonymization

International Journal of Advance Engineering and Research Development. A Survey on Data Mining Methods and its Applications

On Sample Weighted Clustering Algorithm using Euclidean and Mahalanobis Distances

Based on Raymond J. Mooney s slides

A Memetic Heuristic for the Co-clustering Problem

Mining Web Data. Lijun Zhang

REMOVAL OF REDUNDANT AND IRRELEVANT DATA FROM TRAINING DATASETS USING SPEEDY FEATURE SELECTION METHOD

Fuzzy C-means Clustering with Temporal-based Membership Function

Web Information Retrieval using WordNet

Inital Starting Point Analysis for K-Means Clustering: A Case Study

Global Journal of Engineering Science and Research Management

An Intelligent Clustering Algorithm for High Dimensional and Highly Overlapped Photo-Thermal Infrared Imaging Data

Efficiency of k-means and K-Medoids Algorithms for Clustering Arbitrary Data Points

Incremental K-means Clustering Algorithms: A Review

DENSITY BASED AND PARTITION BASED CLUSTERING OF UNCERTAIN DATA BASED ON KL-DIVERGENCE SIMILARITY MEASURE

A Miniature-Based Image Retrieval System

Designing and Building an Automatic Information Retrieval System for Handling the Arabic Data

K-modes Clustering Algorithm for Categorical Data

Mining di Dati Web. Lezione 3 - Clustering and Classification

CHAPTER 4 FUZZY LOGIC, K-MEANS, FUZZY C-MEANS AND BAYESIAN METHODS

An Enhanced K-Medoid Clustering Algorithm

A Survey On Different Text Clustering Techniques For Patent Analysis

Unsupervised Learning

Cluster Analysis. Ying Shen, SSE, Tongji University

Clustering. Chapter 10 in Introduction to statistical learning

CSE 5243 INTRO. TO DATA MINING

A REVIEW ON K-mean ALGORITHM AND IT S DIFFERENT DISTANCE MATRICS

Machine Learning using MapReduce

Hierarchical Clustering Algorithms for Document Datasets

New Approach for K-mean and K-medoids Algorithm

Global Journal of Engineering Science and Research Management

An Approach to Improve Quality of Document Clustering by Word Set Based Documenting Clustering Algorithm

A Fuzzy C-means Clustering Algorithm Based on Pseudo-nearest-neighbor Intervals for Incomplete Data

Cluster analysis of 3D seismic data for oil and gas exploration

Transcription:

IJSRD - International Journal for Scientific Research & Development Vol. 3, Issue 02, 2015 ISSN (online): 2321-0613 An Appropriate Similarity Measure for K-Means Algorithm in Clustering Web Documents S Jaiganesh 1 Dr. P. Jaganathan 2 1,2 Department of Computer Applications 1,2 PSNA College of Engineering and Technology, Dindigul, India Abstract Organizing a large volume of documents into categories through clustering facilitates searching and finding the relevant information on the web easier and quicker. Hence we need more efficient clustering algorithms for organizing large volume of documents. Clustering on large text dataset can be effectively done using partitional clustering algorithms. The K-means algorithm is the most suitable partitional clustering approach for handling large volume of data. K-means clustering algorithm uses a similarity metric that determines the distance from a document to a point that represents a cluster head. This similarity metric plays a vital role in the process of cluster analysis. The usage of suitable similarity metric improves the clustering results. There are varieties of similarity metrics available to find the similarity between any two documents. In this paper, we analyse the performance and effectiveness of these similarity measures in particular to k- means partitional clustering for text document datasets. We use seven text document datasets and five similarity measures namely Euclidean distance, cosine similarity, Jaccard coefficient, Pearson correlation coefficient and Kullback-Leibler Divergence. Based on our experimental study, we conclude that cosine correlation measure is the best suited similarity metric for K-means clustering algorithm. Key words: Partitional Clustering, Cosine Similarity, Euclidean, Jaccard, Pearson, KLD I. INTRODUCTION Clustering is a popular technique in data mining deals with the process of grouping a set of objects into clusters so that objects within the same cluster are similar to each other but are dissimilar to objects in other clusters [1, 6]. Document clustering is an important process that helps in effective organization of documents which leads to retrieve documents quickly and effectively. The clustering techniques are broadly classified into partitioning clustering and hierarchical clustering. The partitioning technique divides the given group of documents into well defined and unique clusters. The hierarchical clustering builds a hierarchy of clusters, showing the relations between individual members and merging clusters of data based on similarity. The partitional clustering technique is the most appropriate for clustering a large document dataset. K- means clustering algorithm is considered as an appropriate partitional clustering approach for clustering a large volume of data [14]. Document clustering groups similar documents into coherent clusters, while documents that are different have separated apart into different clusters. When clustering is applied on web sites, we are usually more interested in clustering the component pages according to the type of information that is presented in the page. Accurate clustering requires a precise definition of the closeness between a pair of objects, in terms of either the pairwised similarity or distance. A wide variety of similarity measures have been used for clustering, such as Euclidean distance, cosine similarity etc. In this work, we have taken five similarity measures namely Euclidean distance, cosine similarity, Jaccard coefficient, Pearson correlation coefficient and Kullback-Leibler Divergence for analysis. The rest of this paper is arranged as follows: The next section describes the document representation for clustering problems. Section 3 discusses the similarity measures and their semantics. Section 4 presents the K- means clustering algorithm and Section 5 explains experiment settings, evaluation approaches, results and analysis. Section 6 concludes and discusses future work. II. DOCUMENT REPRESENTATION FOR CLUSTERING PROBLEM In many of the cluster algorithms, the dataset to be clustered is characterized as a set of vectors X={x 1, x 2,., xn}, where xi corresponds to an object and is referred as feature vector. Every object is characterized with set of well defined and accurate attributes. Vector Space Model (VSM) is a statistical model for representing text document as vectors [8]. There are several ways to model a text document. For example, it can be represented as a bag of words, where words are assumed to appear independently and the order is immaterial. The bag of word model is widely used in information retrieval and text mining [12]. If a term occurs in the document, its value in the vector is non-zero. The term weight value specifies the importance of the term in a document [13]. In the classic vector space model proposed by Salton, Wong and Yang the term specific weights in the document vectors are products of local and global parameters. We understand that the most frequent terms are not necessarily the most informative ones. On the contrary, terms that appears frequently in a small number of documents but rarely in the other documents tend to be more relevant and specific for that particular group of documents, and therefore more useful for finding similar documents. In order to capture these terms and reflect their importance, we transform the basic term frequencies tf(d, t) into the tfidf (term frequency and inversed document frequency) weighting scheme. The model is known as term frequencyinverse document frequency model. Equation (1) gives the weight of term i in document j: (1) Where tf ji refers the number of presence of term i in document j; df ji refers the term frequency in the group of documents; and n refers the total number of documents in the dataset. We measure the degree of similarity of two documents as the correlation between their corresponding vectors, which can be further quantified as the cosine of the All rights reserved by www.ijsrd.com 408

angle between the two vectors. We applied several standard transformations on the basic term vector representation. First, we removed stop words. There are words that are nondescriptive for the topic of a document, such as a, and, are and do. There are 527 universally accepted stop words. Different versatile words were stemmed using Porter s suffix-stripping algorithm, so that words with different endings will be mapped into a single word [11]. For example production, produce, produces and product will be mapped to the stem produce. The underlying assumption is that different morphological variations of words with the same root/stem are thematically similar and should be treated as a single word. We considered the effect of including infrequent terms in the document representation on the overall clustering performance and decided to discard words that appear with less than a given threshold frequency. The rationale by discarding infrequent terms is that in many cases they are not very descriptive about the document s subject and make little contribution to the similarity between two documents. Meanwhile, including rare terms can also introduce noise into the clustering process and make similarity computation more expensive. III. SIMILARITY MEASURES With documents presented as vectors, we measure the degree of similarity of two documents as the correlation between their corresponding vectors, which can be further quantified as the cosine of the angle between the two vector [10]. Before clustering, a similarity distance measure must be determined. The measure reflects the degree of closeness or separation of the target objects and should correspond to the characteristics that are believed to distinguish the clusters embedded in the data [2]. In many cases, these characteristics are dependent on the data or the problem context at hand, and there is no measure that is universally best for all kinds of clustering problems. Thus, a similarity metric has to be defined in any text document analysis. Moreover, choosing an appropriate similarity measure is crucial for cluster analysis, especially for a particular type of clustering algorithms [4]. Therefore, understanding the effectiveness of different measures is of great importance in helping to choose the best one. In general, similarity/distance measures map the distance or similarity between the symbolic descriptions of two objects into a single numeric value, which depends on two factors, the properties of the two objects and the measure itself [5]. In order to make the results of this study comparable to previous research, we include all the measures that were tested in [7] and add another one the averaged Kullback- Leibler divergence. These five measures are discussed below. A. Euclidean Distance: Euclidean distance is a standard metric used for measureing distance between two points and can be easily measured with a ruler in two or three dimensional space. Euclidean distance is widely used in clustering problems, including clustering text documents. It is also the default distance measure used with the K-means algorithm. The Minkowski distances based formula for finding the similarity between document mp and document mj is given by equation (2): ( ) ( ) (2) The Minkowski distance is a metric on Euclidean space which can be considered as a generalization of both Euclidean distance and the Manhattan distance. When the value of n is 2, Euclidean distance(ed) is obtained. Here Normalized Euclidean distance is used as the similarity measure for the documents, mp and mj. The distance measure is calculated using the equation (3): ( ) ( ) (3) where mp and mj are document vectors; dm refers the dimension number of the vector space; mpk and mjk refers the weight values in dimension k for the documents mp and mj. B. Cosine Similarity: Cosine similarity is one of the most popular similarity measure applied to text documents. When documents are represented as term vectors, the similarity of two documents corresponds to the correlation between the vectors. This is quantified as the cosine of the angle between vectors, that is, the so-called cosine similarity. Cosine correlation (CC) is another frequently utilized similarity metric and is given by equation (4): where vectors. ( ) C. Jaccard Coefficient: (4) refers the intersection of the two document The Jaccard coefficient measures similarity as the intersection divided by the union of the objects. For text document, the Jaccard coefficient compares the sum weight of shared terms to the sum weight of terms that are present in either of the two documents but are not the shared terms. The Jaccard similarity coefficient is a statistic used for comparing the similarity and diversity of sample sets. The Jaccard coefficient is calculated using the formula (5): (5) The Jaccard distance, which measures dissimilarity between sample sets, is complementary to the Jaccard coefficient and is obtained by subtracting the Jaccard coefficient from 1, or, equivalently, by dividing the difference of the sizes of the union and the intersection of two sets by the size of the union: D. Pearson Correlation Coefficient: Pearson s correlation coefficient was developed by Karl Pearson from a related idea introduced by Francis Galton in the 1880. It is another measure of the extent to which two vectors are related. There are different forms of the Pearson correlation coefficient formula. Pearson's correlation coefficient when applied to a population is commonly represented by the Greek letter ρ (rho) and may be referred to as the population correlation coefficient or the population Pearson correlation coefficient. The formula for ρ is given below (7): (6) All rights reserved by www.ijsrd.com 409

E. Kullback-Leibler Divergence (KLD): In information theory based clustering, a document is considered as a probability distribution of terms. The similarity of two documents is measured as the distance between the two corresponding probability distributions. The Kullback-Leibler Divergence (KLD), also called the relative entropy, is a widely applied measure for evaluating the differences between two probability distributions. The Kullback Leibler Divergence (also information divergence, information gain, relative entropy, or KLIC; here abbreviated as KLD) is a non-symmetric measure of the difference between two probability distributions P and Q. The KLD of Q from P is defined to be IV. K-MEANS CLUSTERING ALGORITHM We have selected K-means algorithm for performing clustering process. As mentioned previously, partitional clustering algorithms have been recognized to be better suited for handling large document datasets than hierarchical ones, due to their relatively low computational requirements [3,9,15]. The standard K-means algorithm works as follows. Given a set of data objects D and a pre-specified number of clusters k, k data objects are randomly selected to initialize k clusters, each one being the centroid of a cluster. The remaining objects are then assigned to the cluster represented by the nearest or most similar centroid. Next, new centroids are re-computed for each cluster and in turn all documents are re-assigned based on the new centroids. This step iterates until a converged and fixed solution is reached, where all data objects remain in the same cluster after an update of centroids. The generated clustering solutions are locally optimal for the given data set and the initial seeds. Different choices of initial seed sets can result in very different final partitions. The K-means algorithm works with distance measures which basically aims to minimize the within-cluster distances. Therefore, similarity measures do not directly fit into the algorithm, because smaller values indicate dissimilarity. The Euclidean distance and the KLD are distance measures, while the cosine similarity, Jaccard coefficient and Pearson coefficient are similarity measures. We apply a simple transformation to convert the similarity measure to distance values. V. EXPERIMENTS AND RESULTS We have chosen the datasets that have been used commonly for clustering so as to compare and analyse with previous researches. All of these datasets have appropriate preassigned category labels. The rest of this section describes about the datasets, then explains the evaluation measures, and finally presents and analyses the experiment results. A. Datasets: The dataset used for our experiment is widely used in many feature selection techniques, classification and clustering. These datasets differ in terms of document type, number of categories, average category size, and subjects. In order to (7) (8) ensure diversity, the datasets are from different sources, some containing newspaper articles, some containing newsgroup posts, some being web pages and the remaining being academic papers. The number of classes in dataset varies from 4 to 20. As mentioned previously, we removed stop words and applied stemming as described in Section 2, and only the top 2000 words are selected. More specifically, the 20news dataset contained newsgroup articles from 20 newsgroups on a variety of topics including politics, computers, etc. The classic dataset contains abstracts of scientific papers from four sources: CACM, CISI, CRAN and MED. It has been widely used to evaluate information retrieval systems. The hitech dataset consists of San Jose Mercury newspaper articles on six topics computers, electronics, health, medical, research, and technology, and was part of the TREC collection [16]. The tr41 data set is also derived from the TREC-5, TREC-6 and TREC-7 collections. The wap and the webkb datasets both consists of web pages. The wap dataset is from the WebAce project and contains web pages from the Yahoo. The webkb was from the Web Knowledge Base project and contains web pages from several universities about courses, students, staffs, departments, projects and the like. The re0 dataset contains newspaper articles and has been widely used for evaluating clustering algorithms. Data Set Documents Classes Terms Average Class size 20news 18828 20 28553 1217 Classic 7089 4 12009 1774 Hitech 2301 6 13170 384 Re0 1504 13 2886 131 Tr41 878 10 7454 88 Wap 1560 20 8460 78 Webkb 8282 7 20682 1050 Table 1: Summary of Text Dataset VI. EVALUATION We obtained clusters from k-means for each of the datasets taken for study. The number of clusters is set as the same with the number of pre-assigned categories in the data set. The quality of a clustering result was evaluated using two evaluation measures purity and entropy [6, 8]. The purity measure evaluates the coherence of a cluster, that is, the degree to which a cluster contains documents from a single category. Given a particular cluster Ci of size ni, the purity of Ci is formally defined as (9) Purity can be interpreted as the classification rate under the assumption that all samples of the cluster are predicted to be members of the actual dominant class for the cluster. For an ideal cluster, which only contains documents from a single category, its purity value is 1. In general, the higher the purity value, the better the quality of the cluster is. As shown in Table 2, Euclidean distance performs worst and Cosine correlation similarity measure performs better and the best result is achieved by cosine correlation in 5 different dataset out of 7 available. Table 3 shows entropy All rights reserved by www.ijsrd.com 410

results, which indicates that the cosine correlation has lowest purity score on this entire dataset. Eucli Cosine Jaccard Pearson Dataset dean KL Correlat Coefficie Cofficie Dista D ion nt nt nce 20news 0.2 0.62 0.51 0.53 0.41 Classic 0.65 0.97 0.98 0.91 0.91 Hitech 0.31 0.57 0.51 0.56 0.51 Re0 0.61 0.78 0.75 0.78 0.72 Tr41 0.69 0.72 0.73 0.73 0.71 Wap 0.23 0.60 0.63 0.61 0.66 Webkb 0.28 0.68 0.57 0.68 0.61 Table 2: Purity Results Eucli Cosine Jaccard Pearson Data dean KL Correlat Coefficie Cofficie set Dista D ion nt nt nce 20news 0.95 0.49 0.51 0.49 0.54 Classic 0.78 0.25 0.36 0.27 0.3 Hitech 0.92 0.64 0.68 0.65 0.63 Re0 0.6 0.25 0.33 0.26 0.25 Tr41 0.62 0.33 0.33 0.33 0.38 Wap 0.75 0.39 0.4 0.39 0.4 webkb 0.93 0.6 0.74 0.61 0.51 Table 3: Entropy Results The overall entropy value for each measure is shown in Table 3. The overall purity values by all the five similarity measures are very close. Euclidean distance is again proved to be an ineffective metric. The cosine similarity measure tends to outperform other measures. This means that cosine similarity measure is more effective in finding more balanced cluster structures. Considering that the above results all seem very close, in order to test the statistical significance between different solutions, we used the t-test to compare the solutions. First, each solution is represented with a set of clusters and each individual cluster is in turn represented by its centroid vector. Meanwhile, the original dataset is represented as a set of pre-defined categories, and each category is also represented with its centroid. Then both the original dataset and the generated clustering solution are transformed into matrices, with each row being a centroid object and each column is a distinct term from the term set. Finally, the two matrices are tested with the t-test function in the R package.2 The t-test result of any given two matrix shows that there is a true difference between any two matrices, because all the p-values are less than 0.9. VII. CONCLUSION AND FUTURE DIRECTIONS In this work, we found that except for the Euclidean distance measure, the other measures have comparable effectiveness for the partitional text document clustering task. But, cosine correlation coefficient performs slightly better than other measures in producing better clustering solutions. The overall purity values are close and sometimes with only 1% difference. Euclidean distance is proved to be an worst metric. The cosine similarity measure proved that it is better than other measures. We investigate the impact of using different document representation on clustering performance, and combine the different representations with similarity measures. We plan to investigate the effectiveness of these similarity measures with a multi-view clustering approach. In many cases we can view a given document from more than one perspective. For example, web pages intuitively provide at least three views the content text appear in the web page itself, the anchor text of the outgoing links that are embedded in the page and the anchor texts from incoming links. We assume that by dividing the mixed representation up and using the different aspects individually can provide relevant information that is well separated according to its characteristics, therefore benefiting subsequent clustering. REFERENCES [1] Anil K. Jain, Data clustering: 50 years beyond K- means, Pattern Recognition Letters, Vol: 31, pp: 651 666, 2010. [2] Anoop Kumar Jain, Satyam Maheswari, Survey of recent clustering techniques in data mining, International Archive of Applied Science and Technology, Vol: 3, No.2, pp: 68-75, 2012. [3] Giansalvatore Mecca, Salvatore Raunich, Alessandro Pappalardo, A new algorithm for clustering search results, Data and Knowledge Engineering, Vol:62, pp:504-522, 2007. [4] Jean-François Pessiot, Young-Min Kim, Massih R. Amini, Patrick Gallinari., Improving document clustering in a learned concept space, Information Processing and Management, Vol:46, pp:180-192, 2010. [5] Kalyani S, K.S. Swarup., Particle swarm optimization based K-means clustering approach for security assessment in power systems, Expert Systems with Applications,Vol:38, pp:10839 10846, 2011. [6] Kuang Yu Huang, A hybrid particle swarm optimization approach for clustering and classification of datasets, Knowledge-Based Systems, Vol: 24, pp: 420 426, 2011. [7] Leandro dos Santos Coelho, Cezar Augusto Sierakowski, A software tool for teaching of particle swarm optimization fundamentals, Advances in Engineering Software, Vol: 39, pp: 877 887, 2008. [8] Li-Yeh Chuang, Chih-Jen Hsiao, Cheng-Hong Yang., Chaotic particle swarm optimization for data clustering, Expert Systems with Applications, Vol: 38, pp: 14555-14563, 2012. [9] Lin-chih chen, Using a new relational concept to improve the clustering performance of search engines, Information Processing and Management, Vol: 47, No.2, pp: 287-299, 2011. [10] Ming Li, Hang Li, Zhi-Hua Zhou, Semisupervised document retrieval, Information Processing and Management, Vol:45, pp:341-355, 2009. [11] Neepa Shah, Sunita Mahajan, Document Clustering: A detailed Review, International All rights reserved by www.ijsrd.com 411

Journal of Applied Information Systems, Vol:4, No.5, pp:30-39,2012. [12] Ramiz M. Aliguliyev, Clustering of document collection A weighting approach, Expert Systems with Applications, Vol: 36, pp: 7904-7916, 2009. [13] Ruksana Akter, Yoojin Chung., An Evolutionary Approach for Document Clustering, IERI Procedia Vol:4, pp: 370 375, 2013. [14] Saurabh Sharma, Recent Developments in text clustering techniques, International Journal of Computer Applications, Vol:37, No.6, pp: 14-20, 2012. [15] Sunita Bisht, Amit Paul., Document Clustering: A Review, International Journal of Computer Applications, Vol.73, No.11, pp: 26-34, 2013. [16] Tunchan cura, A particle swarm optimization approach to clustering, Expert Systems with Applications, Vol: 39, pp.1582 1588, 2012. [17] Van D. M., Engelbrecht, A. P. Data clustering using particle swarm optimization. Proceedings of IEEE Congress on Evolutionary Computation (CEC 2003), pp. 215-220, 2003. [18] Vishal Gupta, Gurpreet S. Lehal, A Survey of Text Mining Techniques and Applications, Journal of Emerging Technologies in web intelligence, Vol:1, No.1, pp:60-77, 2009. An Appropriate Similarity Measure for K-Means Algorithm in Clustering Web Documents All rights reserved by www.ijsrd.com 412