Optimization of Search Results with Duplicate Page Elimination using Usage Data

Similar documents
Weighted Page Rank Algorithm Based on Number of Visits of Links of Web Page

Web Structure Mining using Link Analysis Algorithms

An Adaptive Approach in Web Search Algorithm

Clean Living: Eliminating Near-Duplicates in Lifetime Personal Storage

EXTRACTION OF RELEVANT WEB PAGES USING DATA MINING

Term-Frequency Inverse-Document Frequency Definition Semantic (TIDS) Based Focused Web Crawler

WEB STRUCTURE MINING USING PAGERANK, IMPROVED PAGERANK AN OVERVIEW

INTRODUCTION. Chapter GENERAL

A New Technique to Optimize User s Browsing Session using Data Mining

An Enhanced Page Ranking Algorithm Based on Weights and Third level Ranking of the Webpages

Efficient Method of Retrieving Digital Library Search Results using Clustering and Time Based Ranking

Design of Query Recommendation System using Clustering and Rank Updater

International Journal of Scientific & Engineering Research Volume 2, Issue 12, December ISSN Web Search Engine

A GEOGRAPHICAL LOCATION INFLUENCED PAGE RANKING TECHNIQUE FOR INFORMATION RETRIEVAL IN SEARCH ENGINE

Reading Time: A Method for Improving the Ranking Scores of Web Pages

IJRIM Volume 2, Issue 2 (February 2012) (ISSN )

Deep Web Content Mining

International Journal of Advance Engineering and Research Development. A Review Paper On Various Web Page Ranking Algorithms In Web Mining

A Framework for Hierarchical Clustering Based Indexing in Search Engines

Design of Query Suggestion System using Search Logs and Query Semantics

Detecting Near-Duplicates in Large-Scale Short Text Databases

A Novel Architecture of Ontology based Semantic Search Engine

Enhanced Retrieval of Web Pages using Improved Page Rank Algorithm

Weighted Page Content Rank for Ordering Web Search Result

Combinatorial Algorithms for Web Search Engines - Three Success Stories

Weighted PageRank using the Rank Improvement

Word Disambiguation in Web Search

INCORPORATING SYNONYMS INTO SNIPPET BASED QUERY RECOMMENDATION SYSTEM

Detection of Distinct URL and Removing DUST Using Multiple Alignments of Sequences

CADIAL Search Engine at INEX

Lecture 17 November 7

INDEXING FOR DOMAIN SPECIFIC HIDDEN WEB

A Framework for Incremental Hidden Web Crawler

AN ADAPTIVE WEB SEARCH SYSTEM BASED ON WEB USAGES MINNG

A Hybrid Page Rank Algorithm: An Efficient Approach

Content Based Smart Crawler For Efficiently Harvesting Deep Web Interface

Ranking Techniques in Search Engines

Leveraging Set Relations in Exact Set Similarity Join

THE WEB SEARCH ENGINE

Analytical survey of Web Page Rank Algorithm

A STUDY OF RANKING ALGORITHM USED BY VARIOUS SEARCH ENGINE

CLUSTERING, TIERED INDEXES AND TERM PROXIMITY WEIGHTING IN TEXT-BASED RETRIEVAL

Improving Suffix Tree Clustering Algorithm for Web Documents

Automatic Query Type Identification Based on Click Through Information

Context Based Web Indexing For Semantic Web

Deep Web Crawling and Mining for Building Advanced Search Application

A HYBRID MODEL FOR DETECTION AND ELIMINATION OF NEAR- DUPLICATES BASED ON WEB PROVENANCE FOR EFFECTIVE WEB SEARCH

A Novel Link and Prospective terms Based Page Ranking Technique

Indexing in Search Engines based on Pipelining Architecture using Single Link HAC

A New Context Based Indexing in Search Engines Using Binary Search Tree

Mining Web Data. Lijun Zhang

Ontology-Based Web Query Classification for Research Paper Searching

IJREAT International Journal of Research in Engineering & Advanced Technology, Volume 1, Issue 5, Oct-Nov, ISSN:

Letter Pair Similarity Classification and URL Ranking Based on Feedback Approach

Web Search Engines: Solutions to Final Exam, Part I December 13, 2004

REDUNDANCY REMOVAL IN WEB SEARCH RESULTS USING RECURSIVE DUPLICATION CHECK ALGORITHM. Pudukkottai, Tamil Nadu, India

Applying the KISS Principle for the CLEF- IP 2010 Prior Art Candidate Patent Search Task

Site Content Analyzer for Analysis of Web Contents and Keyword Density

A SURVEY ON WEB FOCUSED INFORMATION EXTRACTION ALGORITHMS

COMPARATIVE ANALYSIS OF POWER METHOD AND GAUSS-SEIDEL METHOD IN PAGERANK COMPUTATION

A Survey on k-means Clustering Algorithm Using Different Ranking Methods in Data Mining

A Modified Algorithm to Handle Dangling Pages using Hypothetical Node

Inferring User Search for Feedback Sessions

Web Page Classification using FP Growth Algorithm Akansha Garg,Computer Science Department Swami Vivekanad Subharti University,Meerut, India

Self Adjusting Refresh Time Based Architecture for Incremental Web Crawler

IMPROVING INFORMATION RETRIEVAL BASED ON QUERY CLASSIFICATION ALGORITHM

Research Article Combining Pre-fetching and Intelligent Caching Technique (SVM) to Predict Attractive Tourist Places

An Improved Computation of the PageRank Algorithm 1

An Improved PageRank Method based on Genetic Algorithm for Web Search

Retrieval of Web Documents Using a Fuzzy Hierarchical Clustering

Domain Specific Search Engine for Students

Automated Path Ascend Forum Crawling

Survey on Web Structure Mining

From Passages into Elements in XML Retrieval

CHAPTER 4 PROPOSED ARCHITECTURE FOR INCREMENTAL PARALLEL WEBCRAWLER

US Patent 6,658,423. William Pugh

Domain-specific Concept-based Information Retrieval System

arxiv: v1 [cs.lg] 3 Oct 2018

Model for Calculating the Rank of a Web Page

A Novel Categorized Search Strategy using Distributional Clustering Neenu Joseph. M 1, Sudheep Elayidom 2

Support System- Pioneering approach for Web Data Mining

Lecture Notes: Social Networks: Models, Algorithms, and Applications Lecture 28: Apr 26, 2012 Scribes: Mauricio Monsalve and Yamini Mule

Developing Focused Crawlers for Genre Specific Search Engines

A Model for Interactive Web Information Retrieval

Triple Indexing: An Efficient Technique for Fast Phrase Query Evaluation

HYBRIDIZED MODEL FOR EFFICIENT MATCHING AND DATA PREDICTION IN INFORMATION RETRIEVAL

Web Crawling. Jitali Patel 1, Hardik Jethva 2 Dept. of Computer Science and Engineering, Nirma University, Ahmedabad, Gujarat, India

Enhancement in Weighted PageRank Algorithm Using VOL

Automation of URL Discovery and Flattering Mechanism in Live Forum Threads

Correlation Based Feature Selection with Irrelevant Feature Removal

LITERATURE SURVEY ON SEARCH TERM EXTRACTION TECHNIQUE FOR FACET DATA MINING IN CUSTOMER FACING WEBSITE

Design and Implementation of Search Engine Using Vector Space Model for Personalized Search

Proximity Prestige using Incremental Iteration in Page Rank Algorithm

The Anatomy of a Large-Scale Hypertextual Web Search Engine

A New Technique for Ranking Web Pages and Adwords

OPTIMIZED METHOD FOR INDEXING THE HIDDEN WEB DATA

Analysis of Link Algorithms for Web Mining

Automatic Identification of User Goals in Web Search [WWW 05]

DATA MINING - 1DL105, 1DL111

Smart Crawler: A Two-Stage Crawler for Efficiently Harvesting Deep-Web Interfaces

Transcription:

Optimization of Search Results with Page Elimination using Usage Data A. K. Sharma 1, Neelam Duhan 2 1, 2 Department of Computer Engineering, YMCA University of Science & Technology, Faridabad, India 1 Email: ashokkale2@rediffmail.com 2 Email: neelam_duhan@rediffmail.com Abstract The performance and scalability of search engines are greatly affected by the presence of enormous amount of duplicate data on the World Wide Web. The flooded search results containing a large number of identical or near identical web pages affect the search efficiency and seek time of the users to find the desired information within the search results. When navigating through the results, the only information left behind by the users is the trace through the pages they accessed. This data is recorded in the query log files and usually referred to as Web Usage Data. In this paper, a novel technique for optimizing search efficiency by removing duplicate data from search results is being proposed, which utilizes the usage data stored in the query logs. The duplicate data detection is performed by the proposed Data Detection (D3) algorithm, which works offline on the basis of favored user queries found by pre-mining the logs with query clustering. The proposed result optimization technique is supposed to enhance the search engine efficiency and effectiveness to a large scale. In this paper, an approach has been proposed that filters the search results by eliminating duplicate pages detected by the proposed Data Detection (D3) algorithm. The method detects duplicates and near duplicates in an offline mode, while their elimination can be performed online by the search engines. The query logs are pre-mined by applying the query clustering techniques and the discovered query clusters are in turn utilized for finding duplicate pages. The paper has been organized as follows: Section II describes the current research carried out in this area; Section III illustrates the proposed work based on premining the query logs; Section IV shows the performance of proposed work with example scenario and the last section concludes the paper. II. RELATED WORK Index Terms WWW, Search Engine, log, Clustering, PageRank I. INTRODUCTION The Web had been unparalleled in many ways- in scale, in almost complete lack of coordination in its creation, and in the diversity of backgrounds and motives of its participants. One of the essential features, which led to the explosive growth of the web, is the uploading of numerous copies of web documents on the WWW. The tremendous volume of identical or near identical copies of web documents has posed challenges to the performance and scalability of web search engines. data detection is an inherent problem that search engines must have to deal with besides other routine functions. It has been reported that about 10% hosts are mirrored to various extents in a study including 238,000 hosts [1]. The efficient identification of duplicates is a vital issue that has arose from the escalating amount of data and thus needs to be addressed. documents refer not only to completely identical documents but also to nearly identical documents. Though near duplicate documents display striking similarities, they are not bit wise similar [2]. If search engines do not solve the problem of duplicate elimination effectively, generally, in response to a web search, many duplicates may appear in the results. Such duplicates will significantly decrease the perceived relevance of search engines. 248 The notion of Search Result Optimization and Data Detection has been a subject of interest since many years. For optimizing the search results, majority of the present day search engines are using the techniques of one kind or another such as document clustering, page ranking and utilization of user feedback etc. The document clustering methods has been used by many search engines [3] to automatically group the retrieved documents into a list of meaningful categories. Northern Light and Vivisimo are such type of search engines. Ranking methods [4, 5] are also being applied by search engines in order to provide a sorted list of pages to the user based on page relevance. The query logs are considered by many researchers [6] in different aspects of search engines such as building query recommendation systems, inferring semantic concepts or relations from the users queries etc. The proposed duplicate data detection mechanism is also made to utilize query logs, which forms the basis of proposed search result optimization system (Section III). A number of researchers have discussed the problem of finding duplicate data on the web by employing different techniques other than query logs. Some of these methods are described below. An algorithm called Dust Buster [7], for uncovering DUST (Different URLs with Similar Text) was presented in 2007. The method intended to discover rules that transform a given URL to other URLs that are likely to have similar content. Dust Buster employs historical crawl logs or web server logs, instead of probing the page contents, to mine

the dust efficiently. The benefits provided by the method are: increased crawler effectiveness, reduced indexing overhead, and improved quality of popularity statistics such as Page Rank. A technique for estimating the degree of similarity among pairs of documents was presented in 1997, known as shingling [8], which does not rely on any linguistic knowledge other than the ability to tokenize the documents into a list of words. Therefore, it is merely syntactic. In shingling, all word sequences of adjacent words are extracted. If two documents contain the same set of shingles they are considered equivalent and if their sets of shingles appreciably overlap, they are declared similar to different extents. Ilyinsky et al. [9] suggested a method of descriptive words for the definition of near-duplicate documents, which was based on the choice of N words from the index to determine a "signature" of a document. Any search engine based on the inverted index can apply this method. The methods based on "shingles" and signature were compared by the authors. At almost equal accuracy of algorithms, the method in the presence of inverted index was more efficient. An approach called Copy Detection [10] determines the similar web documents, similar sentences and graphically captures the similar sentences in any two web documents. Besides handling wide range of documents, this copy detection approach is applicable to web documents in different subject areas as it does not require static word lists. There are many efficient methods available for search result optimization and duplicate detection [11-15], but a critical look at the available literature indicates that none has used the concept of query logs in duplicate page removal from the search results to optimize and prune the search space and thus, enhance the search efficiency. Moreover, the duplicate detection techniques are not much scalable due to the abundance of web pages on WWW. Therefore, a mechanism needs to be introduced for efficient detection of duplicates and incorporation of the duplicate detection process in the search result optimization. A framework for duplicate data detection and elimination on the basis of mining query logs is proposed in the next section, wherein the offline method uses query log to identify duplicate data and online method filters the search results by eliminating the duplicate pages. The offline method works efficiently without affecting the throughput and performance of search engine. III. PROPOSED RESULT OPTIMIZATION FRAMEWORK A framework for search result optimization has been designed, which detects duplicate pages and eliminates them from the search results (if they happen to appear) as shown in Fig. 1. It has two main modules: the Data Detector and the Eliminator. The Data Detector performs query based duplicate data detection offline by extracting favored or popular user queries by means of query clustering. The information about duplicates is stored in a repository called 249 To & from Search Engine Index Online User Matched Results Processor Eliminator Final Results Access Info Results without s Logs & page Info Data Detector Pages _Doc Database Offline Figure 1. Data Detection & Elimination Framework _Doc Database. The Eliminator works online by removing duplicate pages from the search results by consulting this database. The detailed working of these modules is explained in the next subsections. A. Data Detector This module works on the basis of information extracted from the query logs as shown in Fig. 2. As outlined, initially, a sub-module referred to as Cluster Generator produces clusters of the historical user queries stored in query logs and saves them in an intermediate repository. Then, with the help of Favored query finder, popular user queries are identified from the query clusters. These favored queries are then executed offline and at last duplicate and near duplicate documents are extracted from the retrieved results with the help of Page Extractor and stored to _Doc database. The proposed Data Detector works offline and consists of the following functional components: To & from Processor Logs Page Extractor Page info _Doc Database & Page Info Cluster Generator Favoured Finder Clusters Clusters & Favoured Queries Popular Queries Clusters Figure 2. The Working of Data Detector

1) Cluster Generator This sub-module is used to generate clusters of user queries, which are stored in query logs built by search engines and for this purpose, it uses following two principles: Principle 1 (Based on query contents): If two queries contain same or similar terms, they are supposed to denote the same or similar information needs. Keyword based similarity function is defined as follows: ( ( qi + 1) Sim ( qi, qi + 1) = (1) keyword ( ( qi + 1) where q and q i+1 are two user queries; (q) denotes the number of keywords in a query q. Principle 2 (Based on document clicks): If two queries lead to the selection of the same documents (document clicks), they are considered similar. Clicked URL based similarity function is defined as follows: URL( URL( qi + 1) Sim ( qi, qi + 1) = (2) clickedurl URL( URL( qi + 1) where URL(q) denotes the number of clicked documents with respect to query q. Both principles have been considered important to determine the similarity of queries and thus, any one of them can not be ignored; therefore, a combined measure has been defined to take advantage of both principles as is given below: Sim = λ Sim + μ Sim (3) combined keyword clickedurl where λ and μ are constants with 0 λ, μ 1 and λ+μ =1. A threshold measure (T) for similarity values has been selected to find out the queries to be placed in a particular cluster. The output of Cluster Generator has been shown in Logs Clusters Cluster Generator Cluster1 Cluster2 ClusterN URL URL URL To Favored Finder associated keywords of queries in the set i and their associated clicked URLs in the set URL i. The purpose of generating query clusters is to group similar queries in one cluster. When a new user query is fired, the query terms are matched with keyword set of different clusters and query is assigned to the matched cluster. The corresponding search results of the user query are checked for duplicate elimination with respect to only Algorithm: _Cluster_Generator( ) I/P: Similarity constants λ and μ, Similarity Score threshold T, log with the following fields: 1. IP address of the users. 2. ID: each query is assigned a unique ID 3. : query issued by the user and its tokens. 4. Item rank: if the user clicked on a search result, then rank of the item on which he clicked. 5. Clicked URL: if the user clicked on a search result, the domain position of the URL. O/P: Set of Clusters C= {C1, C2, Cn, each cluster Ci with following information: 1. A cluster-id 2. A collection of related queries. 3. A set of cluster keywords (clust) and cluster URLs (clusturl) of queries present in the cluster. //Start of Algorithm For (each query q i _Log) { Flag [q i ]= 0; // initially each query is unvisited clusturl(= clust(= ClusterId= new (ClusterId); //initialize a ClusterId say C1 For (each query q i ) { If ( Flag (q i ) = =1 ) then continue; //sure that qi is not a part of any existing cluster else //create a new cluster with respect to query q i Assign qi the ClusterId; // Place clicked URLs & tokens of query in respective sets clusturl ( = clusturl ( urls( clust ( = clust ( tokens( For (each query q i+1 ) { // find the similarity of two queries using (3) Find Sim combined (q i, q i+1 ); If ( Sim combined (q i, q i+1 ) > = T) then { Assign q i+1 the same ClusterId as of qi; clusturl ( = clusturl( urls( qi+ 1) ; clust ( = clust ( tokens( qi+ 1) ; Flag [qi]= 1; ClusterId= next (ClusterId) //Initialize the next Id (C2, C3 etc.) //end inner for //end outer for Return the clusters C1, C2, Cn Figure 3. The output of Cluster Generator Fig. 3, where each i th cluster consists of similar queries, the 250 Figure 4. Algorithm for Cluster Generator the matched cluster. The detailed algorithm of cluster generator has been explained in Fig. 4.

2) Favored Finder Once query clusters are formed, the next step is to find a set of favored/popular queries from each cluster. A favored query is one, which occupies a major portion of the search requests in a cluster and it is assumed that, in general, its corresponding search results tend to have more duplicates in search results than disfavored queries. The process of finding favored queries has been shown in Fig. 5. 3) Page Extractor After extracting the popular user queries from their respective query clusters, next step is to find the duplicate pages corresponding to these queries. For this purpose, at the back end, Page Extractor module is made to retrieve a set of search results for each of these queries independently with the help of query processor of search engine and find duplicate page information from this set by applying the proposed Data Detection (D3) algorithm. The process of detecting duplicate documents has been shown in Fig. 6. The original query processor includes a simple GUI that accepts the user queries and returns result pages, but here it is slightly modified to work in the offline mode also so as to extract favored queries from the repository containing the query clusters and respective favored queries. This task is achieved with the help of Extractor, which returns a favored query as a list of terms. The query processor in turn retrieves a ranked list of documents from the index that are deemed relevant to the query. These retrieved documents may include both duplicate and near-duplicate documents and thus these are examined for duplicate detection by the Similarity Calculator, which applies the D3 algorithm to find the similarity score between two documents. D3 uses the following method to find similarity score (SS) between the two documents D i and D j : n N k, ti Nk, tj k = 1 SS( Di, D j ) = (4) length( D ) length( D ) Algorithm: Favored Finder( ) I/P: A cluster Ci of Queries. O/P: True or False. i //Start of Algorithm 1. Queries that are exactly same club them. 2. Make a set of IP addresses corresponding to the queries. 3. For (each query q Cluster) Calculate the Weight of query as: No. of IP addresses which fired the query Wt = Total no. of IP addresses in that cluster If (Wt >= threshold value) then Return True; // is considered as favored query else Return False; j //query is considered as disfavored. Figure 5. Algorithm for Favored Finder Clusters & Favored queries Extractor Page Extractor Similarity Calculator (D3) where t i and t j represent the sets containing tokens/keywords of documents D i and D j respectively, n is the number of common tokens in t i and t j, N k,ti represent the number of occurrences of k th token in D i. The numerator gives the summation of products of term frequencies of common tokens/keywords; e.g. if a common token (say apple) appears 3 and 9 times in D i and D j respectivey, then one summation term in the numerator would be 3 9. Length(D i ) represents the length of the document D i, which can be calculated by (5). N 2 length( D i ) = N k (5) ti k = 1 Pages _Doc Database, Favored Queries Processor Search results with duplicate Pages Fig. 6. Working of Page Extractor Inverted Index where N represent the total number of tokens in D i, The length is calculated by summation of squares of term frequencies of tokens/keywords in the document. The values for Similarity Score (SS) will come out in the range of [0, 1]. The D3 algorithm is described in Fig. 7. It may be noted that if the value of SS is above a predefined threshold value (C), then documents are considered as near duplicates. The Page Extractor stores the returned duplicate page information into the _Doc database along with the ClusterId to which the related query belongs and the PageRank [5] values of pages. An example _Doc repository is shown in Fig. 8. Each document Di is stored with the associated ClusterId and a list S i of identical or near identical documents. The list of duplicate pages S i corresponding to document Di can be accessed via the next field. B. Eliminator This is a very important module of the framework as it performs online optimization of the search results. It consults the _Doc repository maintained by the Page Extractor at the user search time to 251

Algorithm: D3( ) I/P: Set D of N pages wrt a favored query, similarity threshold C O/P: Documents stored in sets S1, S2, S N // Start of Algorithm 1. Perform steps 2 for each D i D. 2. Gather the keywords of document D i from index. 3. Assign a set Si corresponding to each retrieved document D i that will be used to store similar documents. 4. Initially set Si={ D i 5. for (each document D i D) { for (each document D j ) //D j are all other docs except D i { //Calculate Similarity score between Di, & Dj n SS(D i,d j) = i= 1 length of If ( SS(D i, D j )>= C) then Si= Si D j ; 6. Return S1, S2, S N doc D i * length of perform its intended function. The algorithm for this module has been shown in Fig. 9. When a user submits a query on the interface of search engine, query keywords are passed to the Eliminator as well as are given to the Processor used to retrieve the matched result pages from the inverted index of the search engine. The eliminator matches query keywords with set of each query cluster stored in Cluster repository and the maximum matched clusterid is returned. Now, instead of accessing the whole _Doc repository, only the entries with the matched clusterid are accessed for duplicate removal process. If duplicate documents happen to appear in the results of the user query, the issue is which documents should be kept in the optimized results and which ones should be discarded. The tie-breaker rule adopted by the proposed _Doc Database Document ClusterID Next D1 C1 D3 C2 D6 C1 D9 C1 NULL D14 C3 D15 C4. D16 C5. D21 C3.... i, t1 * Figure 7. The D3( ) Algorithm i, t2 doc Dj Document List D2 D6 D20 D16 D56 D120 D245 D2 D20 D21 method is given below, which utilizes the PageRank [16] values of the similar web pages: Out of m duplicate pages with different URLs, only the page with maximum PageRank score will be retained in the final optimized results, while discarding (m-1) duplicates. It may be that, when a user submits such query, whose keywords do not match with any cluster present in clusters & favored queries database, then only the matched results from the search engine index are returned to the interface, while at the same time, the unmatched user query is checked by the Cluster Generator for possible incremental clustering. This new query can either be merged in an existing cluster or assigned a new one. The _Doc database is maintained only for favored queries and their respective clusters as there are maximum chances for the same or similar query to be submitted again by the users. Moreover, considering only favored queries will place less overhead on the normal functioning of a search engine and will also improve the search efficiency in parallel. IV. EXPERIMENTAL RESULTS The results of experiments done for duplicate data detection and elimination are presented in this section. The queries under consideration were related to news and media. Corresponding to each user query, top 100-200 result pages were considered for the experimentation and their associated keywords were stored in MS-Access. The near duplicates have been detected efficiently by the proposed D3() algorithm. The algorithm was able to find Algorithm: _Eliminator( ) I/P: A user query q, _Doc database, N retrieved result pages, Cluster & favored queries database. O/P: Optimized Result pages (<=N) // Start of Algorithm 1. Tokenize the query q and place in set q; 2. Search for the Clusters in query_cluster database, which contain the maximum of q; 3. If (matched cluster found) then GoTo step 4, else GoTo step 9. 4. Search for the entries Si in _Doc repository corresponding to the clusterids found in Step 2; 5. Compare the Search Results with the representative page of each Si ; 6. If (match succeeds with any Si ) then (a). Search for the pages Si in the search results, let it be set Si ; (b). From the representative page of Si & pages Si, Keep the page with highest PageRank in the search results; (c). Filter out the pages with low PageRank from the search results. 7. If any Si still remains to be examined then GoTo Step 5, otherwise GoTo Step 8; 8. Return optimized result pages. 9. Return N result pages without eliminating duplicates Figure 8. Description of _Doc Repository 252 Figure 9. The Eliminator Algorithm

all pairs of web pages, which were duplicated to each other with respect to the predefined threshold values in the range (0.75-0.85). One example from the Google search results is shown, which indicates the presence of duplicates or near duplicate web documents in the results. For the experimental purposes, the query q fired is given below. q: CBI inquiry sought in Nirupama Pathak Murder Case Corresponding to this query, the following three URLs dictated the highest similarity to each other, when threshold value of 0.80 was taken: URL1:http://timesofindia.indiatimes.com/india/Ambiguities-in- Nirupama-autopsy-report-AIIMSexpert/articleshow/5910173.cms URL2:http://kraran.com/nirupama-pathak-autopsy-report-hasambiguities/ URL3:http://beta.thehindu.com/news/national/article425770.ec e Similarity Score between Doc1, Doc2 and Doc3, corresponding to URL1, 2 and 3 respectively, using similarity measure (3) is shown in Table I. The above results indicate that URL1 and URL3 are matched to a large extent as compared to URL1 and URL2 or URL2 and URL3. By considering the above said lower threshold measure; it is concluded that the three are near duplicates to each other. Only one URL is retained in the results while discarding the other two according to their PageRanks or Click_Counts (if no PageRank exists). From the experiments, the resultant performance of search engines in terms of optimization of search results has been shown in Fig. 10. The average values of percentage optimization (reduction) with respect to queries related to different topics is shown in the figure. It may be observed that the queries related to news resulted in pages with up to 5 to 10% pages containing duplicates in the results, therefore removal of such duplicates resulted in optimization (e.g. removal of 8 pages out of 100 resulted in 8% optimization). TABLE I. SIMILARITY SCORE VALUES FOR SAMPLE DOCUMENTS Documents Doc1 Doc2 Doc3 Doc1 1 0.824868 0.868965 Doc2 0.824868 1 0.813696 Doc3 0.868965 0.813696 1 V. CONCLUSION The explosive growth of web has posed challenges in front of search engines to efficiently retrieve relevant and distinct documents in response to user queries. Document retrieval without duplicates has become a major problem nowadays due to the presence of multiple copies of identical and nearly identical content on the web. The 12 10 8 6 4 2 0 News Related Optimization (%age) Entertainment Related Computer H/W Related Mobile Phone Related Figure 10. The Optimization of Results with Elimination proposed framework addresses this problem with the help of query logs maintained by the search engines. Search result optimization on the basis of usage data proves to be novel in the sense that on submitting the future queries, the distinct pages would be presented and search engine would become adaptive to user needs. The main advantage of the approach is the detection of duplicates without affecting the performance of the search engine as it works in an offline mode. As large repositories would be needed to perform offline filtering of search results, therefore, it is performed online to achieve maximum optimization. The search space can be reduced to a large extent by the proposed mechanism. As a future direction, crawlers of the search engines can be made to utilize the information regarding duplicate pages found by the proposed approach, to enhance the crawler efficiency also. REFERENCES [1] Bharat, K., Broder, A.Z, 1999: Mirror, mirror on the Web: A study of host pairs with replicated content. In Proceedings of the 8th International World Wide Web Conference (WWW), pp: 501-512. [2] Xiao, C., Wang, W., Lin, X., Xu Yu, J., 2008. "Efficient Similarity Joins for Near Detection", Proceeding of the 17th international conference on World Wide Web, pp: 131-140. [3] Claudio Carpineto, Stanislaw Osiński, Giovanni Romano, Dawid Weiss. A survey of Web clustering engines, ACM Computing Surveys (CSUR), Volume 41, Issue 3 (July 2009), Article No. 17, ISSN:0360-0300. [4] Neelam Duhan, A. K. Sharma, Komal Kumar Bhatia, 2009. Page Ranking Algorithms: A Survey, Proceedings of IEEE International Advance Computing Conference, (IACC 2009), pp. 2811-2818. [5] Lawrance page and sergey Brin, 1998. The Page Rank Citation ranking: Bringing order to the web. In Proceedings of seventh web conference (WWW 98). [6] R. Baeza-Yates, 2004. usage mining in search engines, Web Mining: Applications and Techniques, Anthony Scime, editor. Idea Group. [7] BarYossef, Z., Keidar, I., Schonfeld, U., 2007. "Do Not Crawl in the DUST: Different URLs with Similar Text", 16th International world Wide Web conference, Alberta, Canada, Data Mining Track, 8-12 May. [8] Broder, A., Glassman, S., Manasse, M., and Zweig, G., 1997. Syntactic Clustering of the Web, In 6th International World Wide Web Conference, pp: 393-404. [9] Ilyinsky, S., Kuzmin, M., Melkov, A., Segalovich, I., 2002. "An efficient method to detect duplicates of Web documents 253

with the use of inverted index", Proceedings of the Eleventh International World Wide Web Conference. [10] Yerra, R., and Yiu Kai, NG., 2005. "A sentence-based Copy Detection Approach for Web Documents", Lecture Notes in Computer Science, Springer Berlin / Heidelberg, Vol. 3613, pp. 557-570. [11] V.A. Narayana, P. Premchand and A. Govardhan, 2009. Effective Detection of Near Web Documents in Web Crawling, Proceedings of International Journal of Computational Intelligence Research, ISSN 0973-1873 Volume 5, Number 1, pp. 83 96. [12] Manku, G. S., Jain, A., Sarma, A. D., 2007. "Detecting nearduplicates for web crawling," Proceedings of the 16th international conference on World Wide Web, pp: 141 150. [13] Henzinger, M., 2006. "Finding near-duplicate web pages: a large-scale evaluation of algorithms," in SIGIR '06: Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval, ACM Press, pp. 284-291. [14] Fetterly, D., Manasse, M., Najork, M., 2003. "On the Evolution of Clusters of Near Web Pages", Proceedings of the First Conference on Latin American Web Congress, pp. 37. [15] Andrei Z. Broder., 2000. "Identifying and Filtering Near- Documents", Proceedings of the 11th Annual Symposium on Combinatorial Pattern Matching. UK: Springer-Verlag, pp.1-10. [16] L. Page, S. Brin, R. Motwani, and T. Winograd, The Pagerank Citation Ranking: Bringing order to the Web. Technical report, Stanford Digital Libraries SIDL-WP-1999-0120, 1999. 254