CHAPTER 5 OPTIMAL CLUSTER-BASED RETRIEVAL

Similar documents
Text Mining Approach for Prediction of Tumor Using Ontology Based Particle Swarm Optimization with Clustering Techniques

Cross-Instance Tuning of Unsupervised Document Clustering Algorithms

Document Clustering using Concept Space and Cosine Similarity Measurement

Modified Particle Swarm Optimization

Improved Clustering of Documents using K-means Algorithm

Information Retrieval and Organisation

CHAPTER 4 AN OPTIMIZED K-MEANS CLUSTERING TECHNIQUE USING BAT ALGORITHM

Particle Swarm Optimization applied to Pattern Recognition

AUTOMATED INFORMATION RETRIEVAL MODEL USING FP GROWTH BASED FUZZY PARTICLE SWARM OPTIMIZATION

CHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES

CHAPTER 3 ASSOCIATON RULE BASED CLUSTERING

BIG Data Meetup EMC, Beer-Sheva. Oshry Ben-Harush Data Scientist, EMC IT. Copyright 2012 EMC Corporation. All rights reserved.

International Journal of Advanced Research in Computer Science and Software Engineering

An improved PID neural network controller for long time delay systems using particle swarm optimization algorithm

β-hill Climbing Technique for the Text Document Clustering

Unsupervised Learning

Gene Clustering & Classification

Semi-supervised Single-label Text Categorization using Centroid-based Classifiers

Improving Tree-Based Classification Rules Using a Particle Swarm Optimization

A Naïve Soft Computing based Approach for Gene Expression Data Analysis

Knowledge Discovery using PSO and DE Techniques

Flat Clustering. Slides are mostly from Hinrich Schütze. March 27, 2017

Hierarchical Co-Clustering Based on Entropy Splitting

Clustering of text documents by implementation of K-means algorithms

Semi-Supervised Clustering with Partial Background Information

Bishnu Prasad Gautam and Dipesh Shrestha

A PSO-based Generic Classifier Design and Weka Implementation Study

Chapter 9. Classification and Clustering

Clustering CS 550: Machine Learning

An Efficient Analysis for High Dimensional Dataset Using K-Means Hybridization with Ant Colony Optimization Algorithm

Interactive Text Mining with Iterative Denoising

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION

Evolutionary Clustering using Frequent Itemsets

Expectation Maximization!

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

A formal concept analysis approach to consensus clustering of multi-experiment expression data

AN EFFICIENT HYBRID APPROACH FOR DATA CLUSTERING USING DYNAMIC K-MEANS ALGORITHM AND FIREFLY ALGORITHM

Introduction to Information Retrieval

CSE 7/5337: Information Retrieval and Web Search Document clustering I (IIR 16)

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

A novel supervised learning algorithm and its use for Spam Detection in Social Bookmarking Systems

Unsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi

Data Clustering. Danushka Bollegala

Proximity Prestige using Incremental Iteration in Page Rank Algorithm

Introduction to Information Retrieval

Application of k-nearest Neighbor on Feature. Tuba Yavuz and H. Altay Guvenir. Bilkent University

Reconfiguration Optimization for Loss Reduction in Distribution Networks using Hybrid PSO algorithm and Fuzzy logic

Unsupervised Learning : Clustering

Chapter 4: Text Clustering

Information Retrieval and Web Search Engines

Information Retrieval and Web Search Engines

COMP 465: Data Mining Still More on Clustering

Cluster Evaluation and Expectation Maximization! adapted from: Doug Downey and Bryan Pardo, Northwestern University

CHAPTER 6 ORTHOGONAL PARTICLE SWARM OPTIMIZATION

INF4820 Algorithms for AI and NLP. Evaluating Classifiers Clustering

Unsupervised Learning and Clustering

INF4820, Algorithms for AI and NLP: Evaluating Classifiers Clustering

CLUSTERING. JELENA JOVANOVIĆ Web:

Minoru SASAKI and Kenji KITA. Department of Information Science & Intelligent Systems. Faculty of Engineering, Tokushima University

A MULTI-SWARM PARTICLE SWARM OPTIMIZATION WITH LOCAL SEARCH ON MULTI-ROBOT SEARCH SYSTEM

Handling Multi Objectives of with Multi Objective Dynamic Particle Swarm Optimization

Text clustering based on a divide and merge strategy

Unsupervised Learning and Clustering

PV211: Introduction to Information Retrieval

Designing of Optimized Combinational Circuits Using Particle Swarm Optimization Algorithm

List of Exercises: Data Mining 1 December 12th, 2015

Introduction to Mobile Robotics

COMPARATIVE ANALYSIS OF PARTICLE SWARM OPTIMIZATION ALGORITHMS FOR TEXT FEATURE SELECTION

Introduction to Information Retrieval

Supervised vs. Unsupervised Learning

Semi-supervised learning

ECLT 5810 Clustering

Best First and Greedy Search Based CFS and Naïve Bayes Algorithms for Hepatitis Diagnosis

Robust Descriptive Statistics Based PSO Algorithm for Image Segmentation

Index Terms PSO, parallel computing, clustering, multiprocessor.

ECLT 5810 Clustering

Clustering. Supervised vs. Unsupervised Learning

Classification of Soil and Vegetation by Fuzzy K-means Classification and Particle Swarm Optimization

Basic Tokenizing, Indexing, and Implementation of Vector-Space Retrieval

Fast Efficient Clustering Algorithm for Balanced Data

Key Instance Detection in Multi-Instance Learning

Chapter 14 Global Search Algorithms

Computer Science 572 Exam Prof. Horowitz Monday, November 27, 2017, 8:00am 9:00am

Unsupervised Data Mining: Clustering. Izabela Moise, Evangelos Pournaras, Dirk Helbing

Constraints in Particle Swarm Optimization of Hidden Markov Models

A Network Intrusion Detection System Architecture Based on Snort and. Computational Intelligence

Behavioral Data Mining. Lecture 18 Clustering

Lecture on Modeling Tools for Clustering & Regression

Hybrid Particle Swarm-Based-Simulated Annealing Optimization Techniques

CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM

Traditional clustering fails if:

THREE PHASE FAULT DIAGNOSIS BASED ON RBF NEURAL NETWORK OPTIMIZED BY PSO ALGORITHM

Data Informatics. Seon Ho Kim, Ph.D.

Improved Performance of Unsupervised Method by Renovated K-Means

Replication on Affinity Propagation: Clustering by Passing Messages Between Data Points

Application of Improved Discrete Particle Swarm Optimization in Logistics Distribution Routing Problem

An Intelligent Clustering Algorithm for High Dimensional and Highly Overlapped Photo-Thermal Infrared Imaging Data

Exploratory Analysis: Clustering

INF4820 Algorithms for AI and NLP. Evaluating Classifiers Clustering

Parallel Sparse K-Means for Document Clustering

Transcription:

85 CHAPTER 5 OPTIMAL CLUSTER-BASED RETRIEVAL 5.1 INTRODUCTION Document clustering can be applied to improve the retrieval process. Fast and high quality document clustering algorithms play an important role in effective navigating, summarizing and organizing information (Cui et al 2005). The clustering methods may improve the results of text mining application. The goal of document clustering is to partition documents into clusters according to their similarities. Major steps of document clustering include tokenization, stemming, elimination of stop words, index term selection, the term document matrix setup and clustering. The Particle Swarm Optimization algorithm is a population based stochastic optimization technique that can be used to find an optimal or near optimal solution to a numerical and qualitative problem (Eberhart and Kennedy 1995). PSO can be successfully applied to a wide range of applications and a survey of the method can be found in Kennedy et al (2001). The clustering problem is an optimization problem that locates the optimal centroids of the clusters. One of the most popular in clustering is K-means algorithm (Niknam et al 2009). The drawback of K-means algorithm is that it may end in local optima. Therefore, it is necessary to use some other global searching algorithm for generating the clusters. The PSO can be used to generate the clusters and it can avoid being trapped in the local optimal solution. In the PSO document clustering algorithm, the multi-dimensional

86 document vector space is modeled as problem space (Cui et al 2005). In Cui and Potok (2005), the document clustering is evaluated using the average distance of document to cluster centroid (ADDC). Hotho et al (2002) have used the approach of applying background knowledge during preprocessing in order to improve clustering results for selection between results. Zhang et al (2007) have discussed various semantic similarity measures of terms including path based similarity measure, information content based similarity measure and feature based similarity measure. Song et al (2008) have discussed how to cluster Deep Web databases semantically. The semantic measure uses ontology and then a hybrid PSO algorithm is provided for deep web databases clustering. Dash et al (2010) have used the hybrid clustering algorithm to find the centroids of a user specified number of clusters, where each cluster groups similar patterns. Killani et al (2010) present a comparative analysis of K-means and PSO clustering performances for text datasets. The result in the work shows the PSO approaches find better solution compared to K-means due to its ability to evaluate many cluster centroids simultaneously in any given time unlike K-means. The clustering problem can be viewed as an optimization problem. In the PSO algorithm, the particle s location represents one solution for the problem. A different problem solution can be generated based on the particle movements to a new location. In each movement of the particle, the velocity and direction will be altered. While most of the clustering procedures perform local searching, the PSO algorithm performs a global searching for solutions. The objective of the PSO clustering algorithm is to discover the proper centroids of clusters for minimizing the intra-cluster distance as well as maximizing the distance between clusters. The PSO algorithm performs a globalized searching for solutions whereas most other partitional clustering procedures perform a localized searching (Cui et al 2005). The optimized

87 clustering model that uses annotation is proposed to inherit the advantages of semantics and the relationship between the terms. The method is used to enhance the effectiveness of document clustering algorithms. The PSO algorithm can be applied on the annotated documents to find the optimal centroids of the clusters. The clustering performance is evaluated using particle swarm optimization. 5.2 FRAMEWORK FOR ONTOLOGY-BASED CLUSTERING Clustering is an important task in information retrieval to improve the precision and recall. The document clustering can be viewed as a utility to assist in document retrieval. Most of the current document clustering approaches work with vector space model and each document is represented by term vector. One major problem with the clustering method is that it does not consider the semantics of the terms. To avoid this problem, the semantic similarity model is introduced. Figure 5.1 shows the architecture of the ontology clustering model. Source Documents GATE Framework Ontologybased Annotation and Indexing Domain Ontology Query Clustering using PSO algorithm Report of Clustered Result Figure 5.1 Ontology-based model for document clustering

88 In contrast to the existing methods, this research approach provides effective clustering by using the semantic similarity that is calculated in document annotation process. The proposed approach is composed of two main parts. Firstly, the document annotation is done using ontology and secondly, the optimized clustering approach is used to improve the relevance. The novelty of this approach resides in the annotation and in applying PSO to retrieve optimized result. In the traditional retrieval system, keyword method cannot meet the need of user s requirements. The clustering techniques group the documents into unsupervised cluster based on the features of the documents. The document annotation model is utilized for the clustering of the documents. The main features of the model proposed by this research, in comparison with the models such as boolean, statistical and probabilistic are: 1. Combining of the conceptual features of documents and to develop an effective annotation 2. Using semantic similarity score to improve the concept relevance 3. Using optimization algorithm to improve the clustering performance. 5.3 REPRESENTATION OF PSO CLUSTERING In the PSO system, particles move through the search space which allows the particles to explore more information space and better the chance of finding global optima. One particle s position in the swarm represents one possible solution for clustering the document collection. Therefore, a swarm represents a number of candidate clustering solutions for the document collection (Cui and Potok, 2005). Each particle maintains a matrix P1 g = (cc g1, cc g2, cc gg,.., cc gk ), where cc gk is the cluster centroid of k th cluster. The cc gh is

89 represented as position vector x gh. x gh is the h th element position of the g th particle. The particle in the swarm has its own position and velocity. The movement of the particle is directed by the velocity. The term v gh represents the h th element of the velocity vector of the g th particle. For every generation, the particle s new location is computed by adding the particle s current velocity V-vector to its location X-vector. In Equation (5.1), the random values rand 1 and rand 2 are used for a wide search. The factor is the inertia random value. The values of c 1 and c 2 control the weight in deciding the particle s next movement. v v c rand ( Ppbest x ) c rand ( Pgbest x ) (5.1) gh gh 1 1 gh 2 2 gh xgh xgh vgh (5.2) At every generation, the particle s new location is computed by adding the particle s current velocity, v gh, to its location, x gh using Equation (5.1) and Equation (5.2) and thus the velocity of the particle is updated (Kennedy et al 2001). Each particle stores the personal and global best position in the search space (Kennedy and Eberhart, 1995). The personal best position of g th particle is the best position visited by the particle so far and is denoted as Ppbest g (t). Let f be the objective function and the personal best of particle at time step t is updated through Equation (5.3). Ppbestg ( t) if f ( xg ( t 1) f ( Ppbestg ( t)) Ppbestg ( t 1) xg ( t 1) if f ( xg ( t 1) f ( Ppbestg ( t)) (5.3) The global best solution is represented as Pgbest. The best particle is determined from the entire swarm by selecting the personal best position. The global best is updated based on the best known position of the swarm using Equation (5.4).

90 Pgbest( t) Ppbest, Ppbest,... Ppbest min f ( Ppbest ( t),... Ppbest ( t) 0 1 k 0 (5.4) The PSO clustering algorithm needs a fitness function to evaluate each particle performance. Each particle maintains a matrix with the cluster centroid vectors. The inertia weight changes the momentum of particles to avoid the stagnation of particles at the local optima. The initial iteration can be a global searching stage. After several iterations, the particle s velocity will reduce and the searches will gradually reduce while the particle will approach the optimal solution. k 5.4 CLUSTERING USING PSO ALGORITHM The PSO algorithm is given as follows: Step 1: Randomly initialize the particle velocity and particle position. The K cluster centroids are randomly selected for each particle. Step 2: Evaluate the fitness for each particle for the initial values Step 3: Preserve the document cluster structure optimally. The cluster quality can be measured using within cluster, between-cluster and mixture scatter matrices (Howland and Park, 2004) and it is given in Equation (5.5), Equation (5.6) and Equation (5.7) respectively. The cluster quality is high if the cluster is tightly grouped, but well separated from the other clusters. Equation (5.5) defines the withincluster and the Equation (5.6) defines the similarity measure between clusters. The function preserves the cluster structure and maximizes the Equation (5.7). k T S d cc d cc, (5.5) w j i j i i 1 j N i

91 k T, (5.6) S c gc c gc b i i i 1 j N n i m j j j 1 T S ( d gc) d gc (5.7) where d j is the document that belongs to the cluster cc i. cc i is the i th cluster centroid, gc is the global cluster centroid, K is the number of clusters and N i is the number of documents in the cluster cc i. S w is the within cluster, S b is between clustering and S m is the mixture scatter matrices which is sum of S w and S b. The fitness value of the particle is evaluated using S b. The goal of clustering is to attain high intra-cluster similarity and low inter-cluster similarity. Step 4: Update the personal best position using Equation (5.3) Step 5: Step 6: Step 7: Step 8: Step 9: Update the global best position among all the individuals using Equation (5.4) Apply velocity update for each dimension of the particle using Equation (5.1) Update the position and generate new particle s location using Equation (5.2) Repeat steps 2 8 until a stopping criteria, such as a sufficiently good solution is discovered or a maximum number of generations is completed. The individual particle that scores the best fitness value in the population is taken as optimal solution. The relevance of documents is evaluated

92 5.5 EXPERIMENTAL RESULT 5.5.1 Data set Description The NewsGroups data set is used for text clustering. The Mini News Group which is the subset of 20 NewsGroups is used as a data set and it has a collection of about 2000 documents across 20 different newsgroups from Usenet. Each category consists of 1,000 documents assigned to it. Each newsgroup is stored in a subdirectory, with each article stored as a separate file. Some of the newsgroups are closely related with each other while some are highly unrelated. For the purpose of experimentation, clustering is done using upto 10 groups. Using the approach of Jing et al (2006), the four datasets are extracted from the newsgroup datasets. The datasets D1 and D2 contain categories of different topics. The datasets D3, D4 consist of categories of similar topics collected from the collection. Table 5.1 and Table 5.2 show the different topics organized into 20 newsgroups. Table 5.1 describes the dataset used for evaluation and Table 5.2 shows the topics of the newsgroups. The documents are annotated using KIM Ontology (KIMO) (Popov et al 2003) and it is stored using the GATE tool (Cunningham 2002). Table 5.1 Datasets used for evaluation Dataset No.of Document No. of Classes No.of Unique Terms 20Newsgroup 18,828 20 18,330

93 Table 5.2 Dataset Description comp.graphics comp.os.mswindows.mis comp.sys.ibm.pc.hardwar comp.sys.mac.hardware comp.windows.x misc.forsale rec.autos rec.motorcycles rec.sport.baseball rec.sport.hockey talk.politics.misc talk.politics.guns talk.politics.midea sci.crypt sci.electronics sci.med sci.space talk.religion.misc alt.atheism soc.religion.christian 5.5.2 Cluster Evaluation 20 random particles are generated. The fitness value is calculated using the distance between the cluster centroid and the documents that are clustered. The fitness values are recorded for ten simulations. In the PSO module, the inertia weight initially set as 0.95 is chosen and the acceleration coefficient constants c1 and c2 are set to 2. If the fitness value is fixed, then it indicates the optimal solution. For each simulation, the initial centroid vectors of each particle are randomly selected. A set of 20 queries was prepared manually for comparative performance measurement. After 100 iterations the same cluster solution is obtained. The F-measure values are the average of 20 runs. The parameters and their values are shown in Table 5.3 Experimental results show that the PSO clustering algorithm using ontology performs better than the PSO clustering algorithms without using ontology. To cluster the large document data sets, PSO requires more iteration to find the optimal solution. As shown in Figure 5.2, the PSO approach generates the clustering result with the lowest fitness value for all four datasets using the euclidian similarity metric and the ontology-based similarity metric.

94 Table 5.3 PSO Parameters and their values Parameter Value No. of clusters 10 No. of Particles 20 Maximum no. of Iterations c1 2 c2 2 100 0.95 rand1 and rand2 Uniformly distributed random variable between 0 and 1. Fitness 14 12 10 8 6 4 2 PSO without ontology PSO with ontology 0 0 5 10 15 20 25 30 35 40 45 50 55 60 Iteration Figure 5.2 Fitness curve comparisons

95 The convergence of PSO depends on the particle s best known position and the swarm's best known global position. If the global position remains constant throughout the optimization process then the algorithm converges to the optimal. For a large data set, the PSO requires more iteration before optimal clustering. The PSO clustering algorithm can generate more compact clustering results. The PSO approach in ontology-based VSM has improvements compared to the results of the VSM-based PSO. However, when the similarity metric is changed to the ontology-based metric, the PSO algorithm has a better performance. Table 5.4 Performance comparison of K-Means and PSO Dataset1 Dataset 2 Dataset3 Dataset 4 Fitness Value K-Means PSO Euclidian 8.11009 6.38039 Ontology 10.41271 8.14551 Euclidian 6.25172 4.51753 Ontology 9.57786 7.21153 Euclidian 4.14896 2.25961 Ontology 5.71146 4.00555 Euclidian 8.62794 6.37872 Ontology 12.8927 9.5379 The convergence behavior of K-means and PSO algorithm are given in Table 5.4. Since 100 iterations is not enough for the PSO algorithm to converge to an optimal solution, the result values in the Table 5.4 indicate that the PSO approach have improvements compared to the results of the K-means approach when using the euclidian similarity metric. However, when the similarity metric is changed to the ontology similarity metric, the PSO algorithm has a better performance than the K-means algorithm.

96 5.5.3 Discussion on the Clustering Quality For clustering, two measures of cluster goodness or quality are used. One type of measure allows to compare the different sets of clusters without reference to external knowledge and is called an internal quality measure. The other type of measures permits an evaluation of how well the clustering is working by comparing the groups produced by the clustering techniques to known classes. This type of measure is called an external quality measure. The commonly measured clustering quality F-measure and entropy are used to evaluate the quality of the clusters produced by the clustering algorithm. Entropy is a commonly used external validation measure for clustering (Zhao and Karypis 2004). Entropy measures the purity of the clusters with respect to the given class labels. For each cluster j, compute the probability pij that a member of cluster belongs to class i. The E j entropy of cluster j is then calculated using the below formula k E p log( p ) (5.8) j j 1 ij ij The total entropy is E and the total entropy for a set of clusters is computed as the weighted sum of the entropies of each cluster, as shown in Equation (5.9) E K n1 j Ej (5.9) n j 1 where n1 j is the size of cluster j, K is the number of clusters, and n is the total number of data points. The entropy considers the overall distribution of all the categories in a given cluster. In general, the smaller the entropy value, the better the quality of the cluster is. At each step of the iteration process, the entropy value is given in Figure 5.3.

97 Entropy 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 VSM VSM+ Ontology 0 0 5 10 15 20 25 30 Iteration Figure 5.3 Comparison of clustering quality using entropy Using the suggestions of Manning et al (2008), true positive, true negative, false positive and false negative are calculated. Two similar documents in the same cluster are true positive (TP) and two similar documents in different clusters are true negative (TN). The decision of assigning two dissimilar documents to the same cluster is false positive (FP) and two similar documents to different clusters is false negative (FN). Precision (P) and Recall(R) are two popular measures for clustering. The higher precision is obtained at the low recall value. The F-measure (F) is used to evaluate the performance measure of the model. The F-measure combines Precision and Recall, the two most widely used measures. A higher F- measure indicates better performance. The Precision and Recall are given in Equation (5.10). The F-measure is evaluated using the Equation (5.11).

98 The cluster quality for the four data sets from newsgroups is given in Figure 5.4. For different data sets in 20 Newsgroups the F-measure performance of ontology-based VSM increases greatly when compared to the VSM method. It shows consistent performance in F- measure in ontologybased VSM and is still the best among these methods. The aim of the algorithm is to maximize the F-measure. The higher the overall F-measure, the better the clustering accuracy achieved. The semantic similarity measure outperforms the keyword-based similarity measure model. P TP TP FP (5.10) TP R TP FN (5.11) F 2 P R P R (5.12) As shown in Table 5.5, the ontology-based method produces better results. The traditional VSM with TFIDF and VSM with ontology are compared. It is obvious that the VSM model with ontology similarity plays an important role in accurately judging the relation between documents.

99 1 0.9 0.8 0.7 F- Measure 0.6 0.5 0.4 VSM VSM+Ontology 0.3 0.2 0.1 0 D1 D2 D3 D4 Data Set Figure 5.4 Comparison of clustering quality Table 5.5 Performance evaluation of document representation model Dataset VSM VSM+Onology F-measure Entropy F-measure Entropy D1 0.976 0.089 0.986 0.088 D2 0.847 0.058 0.858 0.052 D3 0.758 0.574 0.776 0.563 D4 0.815 0.396 0.834 0.336 The mean relevance for the top 20 retrieved documents using ontology is 3.1 and the same using the keyword based search is 2.2. VSM+Ontology obtains the best clustering performance in terms of the results of Precision and Recall. Evaluation results on the chosen corpus and the annotation ontology have shown improvements of retrieval performance

100 compared to simple keyword matching approaches. In the traditional VSM the background knowledge is not considered in the term similarity. The performance of ontology-based VSM shows a better improvement than traditional VSM. The clustering results, produced by the semantic similarity approach, have a higher quality than those produced by a keyword based similarity approach. The evaluations results indicate that the semantic annotation, indexing and retrieval approach improve the performances of retrieval not just in terms of Recall, but also in terms of Precision. Based on the clustering results of Jing et al (2006), the clustering quality is compared with the standard K-means and PSO in terms of the ontology distance measure. The traditional VSM is compared with Frequency-based Term Similarity (FBTS) method and Ontology-based FBTS. The standard K-means and PSO on the traditional VSM are compared for their relative improvements on the clustering quality. RIFS is relative improvement in F-measure and RIEN is relative improvement in entropy measure. Table 5.6 shows the relative improvement in F-measure and Entropy. Table 5.6 Relative improvement of F Score and Entropy Methods Measure D1 D2 D3 D4 K-Means PSO RIFS(%) RIEN(%) RIFS(%) RIEN(%) 4.80 5.71 4.58 16.89 7.35 9.89 7.20 18.19 0.88 4.10 0.89 5.45 4.91 13.12 5.3 19.12 Several observations from the comparison between these methods are worth discussing: 1. Our ontology-based VSM clustering results are more accurate than the traditional VSM clustering.

101 2. The semantic similarity performs consistently and significantly better than the keyword-based term frequency methods and achieves the best performance in all experiments. 3. The PSO clustering approach combined with semantic similarity performs significantly the best among the traditional clustering methods in all the experiments of this research. 5.6 SUMMARY The proposed model improves the text clustering model by combining the annotation weights. The aim of the current research work is to provide qualitative improvement over VSM-based search by using ontology. The document clustering is done through Particle Swarm Optimization. A PSO-based ontology model of clustering knowledge documents is presented and compared to the traditional vector space model. The proposed ontology-based framework provides improved performance and better clustering compared to the traditional vector space model. It also overcomes the problems existing in the vector space model commonly used for clustering. The clustering result based on semantic similarity has higher fitness values than those based on the traditional vector space model. It is worth pointing out that the above observations are made by combining semantic similarity and PSO clustering in terms of F-measure.