A NEW PERFORMANCE EVALUATION TECHNIQUE FOR WEB INFORMATION RETRIEVAL SYSTEMS

Similar documents
ARCHITECTURE AND IMPLEMENTATION OF A NEW USER INTERFACE FOR INTERNET SEARCH ENGINES


Query Modifications Patterns During Web Searching

Automatic New Topic Identification in Search Engine Transaction Log Using Goal Programming

Enabling Users to Visually Evaluate the Effectiveness of Different Search Queries or Engines

A Model for Interactive Web Information Retrieval

Mining the Query Logs of a Chinese Web Search Engine for Character Usage Analysis

An Analysis of Document Viewing Patterns of Web Search Engine Users

Using Clusters on the Vivisimo Web Search Engine

Repeat Visits to Vivisimo.com: Implications for Successive Web Searching

CHAPTER THREE INFORMATION RETRIEVAL SYSTEM

WEIGHTING QUERY TERMS USING WORDNET ONTOLOGY

RMIT University at TREC 2006: Terabyte Track

A User Study on Features Supporting Subjective Relevance for Information Retrieval Interfaces

Enhancing Cluster Quality by Using User Browsing Time

Enhancing Cluster Quality by Using User Browsing Time

Parallel Algorithms on Clusters of Multicores: Comparing Message Passing vs Hybrid Programming

Designing and Building an Automatic Information Retrieval System for Handling the Arabic Data

6.2 DATA DISTRIBUTION AND EXPERIMENT DETAILS

Performance Evaluation of XHTML encoding and compression

Web document summarisation: a task-oriented evaluation

Improving Range Query Performance on Historic Web Page Data

Multicast Transport Protocol Analysis: Self-Similar Sources *

BEx Front end Performance

Improving object cache performance through selective placement

IJREAT International Journal of Research in Engineering & Advanced Technology, Volume 1, Issue 5, Oct-Nov, ISSN:

Domain Specific Search Engine for Students

VIDEO SEARCHING AND BROWSING USING VIEWFINDER

Contents Overview of the Compression Server White Paper... 5 Business Problem... 7

QLIKVIEW SCALABILITY BENCHMARK WHITE PAPER

IMPROVING THE RELEVANCY OF DOCUMENT SEARCH USING THE MULTI-TERM ADJACENCY KEYWORD-ORDER MODEL

Examining the Authority and Ranking Effects as the result list depth used in data fusion is varied

A Query-Level Examination of End User Searching Behaviour on the Excite Search Engine. Dietmar Wolfram

A JAVA META-REGISTRY FOR REMOTE SERVICE OBJECTS

Load Balancing Algorithm over a Distributed Cloud Network

Dietmar Wolfram University of Wisconsin--Milwaukee. Abstract

Designing Issues For Distributed Computing System: An Empirical View

RETRACTED ARTICLE. Web-Based Data Mining in System Design and Implementation. Open Access. Jianhu Gong 1* and Jianzhi Gong 2

Integrating VVVVVV Caches and Search Engines*

Design and Implementation of Search Engine Using Vector Space Model for Personalized Search

SEDA: An Architecture for Well-Conditioned, Scalable Internet Services

Automatic Query Type Identification Based on Click Through Information


Speed and Accuracy using Four Boolean Query Systems

Efficient World-Wide-Web Information Gathering. Tian Fanjiang Wang Xidong Wang Dingxing

A PRELIMINARY STUDY ON THE EXTRACTION OF SOCIO-TOPICAL WEB KEYWORDS

Internet Usage Transaction Log Studies: The Next Generation

WHAT HAPPENS IF WE SWITCH THE DEFAULT LANGUAGE OF A WEBSITE?

A Capacity Planning Methodology for Distributed E-Commerce Applications

Spoken Document Retrieval (SDR) for Broadcast News in Indian Languages

Tuning WebHound 4.0 and SAS 8.2 for Enterprise Windows Systems James R. Lebak, Unisys Corporation, Malvern, PA

Revisiting Join Site Selection in Distributed Database Systems

Compressing and Decoding Term Statistics Time Series

Deep Web Content Mining

Configuration Management for Component-based Systems

This is a repository copy of A Rule Chaining Architecture Using a Correlation Matrix Memory.

I. INTRODUCTION FACTORS RELATED TO PERFORMANCE ANALYSIS

Design and Implementation of A P2P Cooperative Proxy Cache System

Flash Drive Emulation

A BFS-BASED SIMILAR CONFERENCE RETRIEVAL FRAMEWORK

Ch. 7: Benchmarks and Performance Tests

DATA MINING II - 1DL460. Spring 2014"

QUERY EXPANSION USING WORDNET WITH A LOGICAL MODEL OF INFORMATION RETRIEVAL

ADAPTIVE AND DYNAMIC LOAD BALANCING METHODOLOGIES FOR DISTRIBUTED ENVIRONMENT

Anatomy of a Semantic Virus

GENERATING SUPPLEMENTARY INDEX RECORDS USING MORPHOLOGICAL ANALYSIS FOR HIGH-SPEED PARTIAL MATCHING ABSTRACT

A Tree-based Inverted File for Fast Ranked-Document Retrieval

Information Retrieval

COVER SHEET. Accessed from Copyright 2003 Elsevier.

ON NEW STRATEGY FOR PRIORITISING THE SELECTED FLOW IN QUEUING SYSTEM

code pattern analysis of object-oriented programming languages

Inverted Indexes. Indexing and Searching, Modern Information Retrieval, Addison Wesley, 2010 p. 5

Equivalence Detection Using Parse-tree Normalization for Math Search

International Jmynal of Intellectual Advancements and Research in Engineering Computations

Review of. Amanda Spink. and her work in. Web Searching and Retrieval,

A Firewall Architecture to Enhance Performance of Enterprise Network

Subjective Relevance: Implications on Interface Design for Information Retrieval Systems

Adaptable and Adaptive Web Information Systems. Lecture 1: Introduction

A Model for Information Retrieval Agent System Based on Keywords Distribution

STRUCTURE-BASED QUERY EXPANSION FOR XML SEARCH ENGINE

Okapi in TIPS: The Changing Context of Information Retrieval

Overview of the INEX 2009 Link the Wiki Track

A Visualization Program for Subset Sum Instances

Analyzing Web Multimedia Query Reformulation Behavior

SCALING UP VS. SCALING OUT IN A QLIKVIEW ENVIRONMENT

CRAWLING THE CLIENT-SIDE HIDDEN WEB

International Journal of Advance Foundation and Research in Science & Engineering (IJAFRSE) Volume 1, Issue 2, July 2014.

Information Retrieval Spring Web retrieval

Data Mining with Oracle 10g using Clustering and Classification Algorithms Nhamo Mdzingwa September 25, 2005

Evaluation of Performance of Cooperative Web Caching with Web Polygraph

WebSphere Application Server 6.1 Base Performance September WebSphere Application Server 6.1 Base Performance

Switched FC-AL: An Arbitrated Loop Attachment for Fibre Channel Switches

Efficient Queuing Architecture for a Buffered Crossbar Switch

Spemmet - A Tool for Modeling Software Processes with SPEM

IJMIE Volume 2, Issue 9 ISSN:

Software re-use assessment for quality M. Ramachandran School of Computing and Mathematical Sciences, Jo/m Moores C/mrerszZ?/,

Capturing and Formalizing SAF Availability Management Framework Configuration Requirements

IMPROVING THE DATA COLLECTION RATE IN WIRELESS SENSOR NETWORKS BY USING THE MOBILE RELAYS

Performance Extrapolation for Load Testing Results of Mixture of Applications

CLUSTERING, TIERED INDEXES AND TERM PROXIMITY WEIGHTING IN TEXT-BASED RETRIEVAL

Transcription:

A NEW PERFORMANCE EVALUATION TECHNIQUE FOR WEB INFORMATION RETRIEVAL SYSTEMS Fidel Cacheda, Francisco Puentes, Victor Carneiro Department of Information and Communications Technologies, University of A Coruña Facultad de Informática, Campus de Elviña s/n, 15.71 A Coruña, Spain* {fidel, fpuentes, vicar}@udc.es ABSTRACT The performance evaluation of an information retrieval system is a decisive aspect for the measure of the improvements in search technology. Our work intends to provide a framework to compare and contrast the performance of search engines in a research environment. In this way, we have designed and developed USim, a tool for the performance evaluation of Web IR systems based on the simulation of users behavior. This simulation tool contributes in the performance evaluation process in two ways: estimating the saturation threshold of the system and in the comparison of different search algorithms or engines. The latter point is the most interesting because, as we demonstrated, the comparison using different workload environments will achieve more accurate results (avoiding erroneous conclusions derived from ideal environments). From a general point of view, USim intends to be an approximation to some new performance evaluation techniques specifically developed for the Internet search engines. KEYWORDS Web Information Retrieval, performance evaluation, simulation. 1. INTRODUCTION With the exponential increase of the Web there has also been a growing interest in the study of a variety of topics related to the use of the Web. There is a special interest in finding patterns in and computing statistics of Internet search engines users and some articles that analyze the query logs of several commercial search engines have been published (Cacheda and Viña, 21), (Jansen et al., 1998), (Kirsch, 1998), (Silverstein et al., 1999) and (Spink et al., 22). These studies examine the search process developed by Web users: terms per query, top terms, operators and modifiers, sessions, etc. Moreover, the performance evaluation of an Information Retrieval (IR) system is a decisive aspect for the measure of the improvements in search technology, as it is derived from the Text Retrieval Conference (TREC) and the WEB-TREC conferences. Zobel, Moffat and Ramamohanarao (1996) describe the guidelines for evaluating the performance and comparing several indexing and retrieval techniques. The main criteria for the comparison are the following ones: the scalability of the system, the response time for the search process (which is perhaps the single crucial test for an indexing scheme), disk space used by the index data structures, CPU time, disk traffic and memory requirements. Nevertheless, the response time is not an easily estimable parameter because it is the aggregate of many other parameters. The same problem arises in the Web IR systems, and gets complicated because these systems must operate under different workload situations, especially with a high number of concurrent users. In fact, in the WEB-TREC one of the measures to obtain is the response time of each request sent to the system (Hawking et al., 1999). However, the response times are computed in an ideal situation: without any workload on the system. This work is focused on the development and testing of search engines in a research environment, in order to provide a framework to compare and contrast the performance obtained after a change in the IR system. With the changes referring to any part of the search engine: the index organization, the search algorithm, the system architecture or the system configuration. For this purpose, we have designed and developed USim, a tool for the performance evaluation of Web IR systems based on the simulation of users behavior, in order to compare more accurately the performance

of different search systems. Moreover, we tested this performance evaluation tool in a real environment, in two ways: estimating the saturation threshold of the system and comparing the performance of different search algorithms or engines. This paper is structured as follows. It starts with an overview of the related research. Next we describe USim, the simulation tool proposed and the next section details the results obtained in the performance evaluation of Web IR systems. Finally the main conclusions are exposed. 2. RELATED STUDIES Recently there have been some studies that examine the behavior of Web search users while they are using an Internet search engine or a Web directory. The first study was performed by Kirsh (1998) who presented some search statistics of Infoseek usage. A bit later, Jansen et al. (1998) presented a study of queries, sessions and searched terms obtained in the query logs of Excite. Spink et al. (22) analyze the changes in the search topics for Excite through several years. Silverstein et al. (1999) examined a very large number of queries taken from AltaVista logs, studying not only the queries but also the co-ocurrences among them. Cacheda and Viña (21) investigated the queries, categories browsed and the documents retrieved by the users of a Web directory. These studies show that most Web users enter few queries consisting of few search terms, have difficulty using effective keyword or Boolean queries and conduct little query reformulation. Also, Jansen and Pooch (21) provide a framework for the development and comparison of future query log analysis, comparing the searching characteristics of Web users with users of other systems. Latterly Spink and Ozmultu (22) analyzed the characteristics of question format web queries and Ozmultu, Spink and Ozmultu (22) explored a sampling technique for the correct statistical analysis of large data sets of web queries. On the other side, the importance of the performance evaluation is well known and it is fundamental to obtain the effectiveness and response time measures of the IR system. For the traditional IR systems, this evaluation is performed using the methodology created in the TREC evaluation program undertaken by the US National Institute of Standards and Technology (TREC, 24). In the case of the Web IR systems, the WEB-TREC was created allowing the measurement of speed (using the response time) and effectiveness (using the precision and recall parameters). But the Web IR systems must operate under different workload situations, especially with a high number of concurrent users. Consequently, it would be quite important to measure the response time of a Web IR system under different workload situations. Clearly, a search engine cannot be put into production before being evaluated and so, no real workload situations (using real users) could be used. Therefore, it is fundamental to obtain a user s profile that could be simulated, by means of the query logs analysis of a Web IR system, to later examine the simulation of different workloads (as a different number of requests) and its effect on the response time of a search engine, in order to improve the performance evaluation methodology. The user s profile is based on the work by Cacheda and Viña (21), where two basic conclusions are obtained for the simulation of the users behavior: - The searches, categories browsed and documents visited fit an Exponential distribution, with the mean variable through the time. - There is a linear relationship between the number of searches, categories browsed and documents visited in a period of time. 3. USIM: A PERFORMANCE EVALUATION TOOL USim (Users Simulator) is an application that simulates the users behavior while using a search engine, using the results obtained in some previous works. It has been designed to operate with any type of search engine using the HTTP protocol. USim will send multiple requests to an IR system in the same way that a group of users would have done. Three types of requests are supported: searches, browsed categories and visited documents. The requests will be sent following an Exponential distribution, as derived from the previous section and the number of requests per minute is defined as a parameter. In this way, USim can generate different supervised workload environments over an IR system.

USim was designed and implemented for research purposes in order to evaluate the performance of a Web IR system in a LAN, before being available on the Web. This guarantees that the network latency is negligible and so, the response times obtained only measure the search engine response times. 3.1 Design and implementation USim was designed using the object oriented methodology and completely developed in the object oriented language Java, in order to build a multiplatform application and facilitate its operation in any environment. The simulation tool is composed of three main modules that operate concurrently, associated with the three types of requests: searches, categories and documents. The three types of requests are generated in a similar way. Each type of request is managed by its own process, and at the same time, each request is processed independently by its own thread. As described in the previous section, the time between two consecutive requests fits an Exponential distribution and the simulation is done using the inversion method over the distribution function of the Exponential distribution. Then, a request is generated, submitted to the search system and then the response is processed when it is received. The generation of the request is handled in different ways depending on the type of the request. In the case of the searches a whole query must be created. Starting with the search string, the analysis carried out by Cacheda and Viña (21) showed that the search strings examined did not fit the Zipf s law (Zipf, 1949), confirming the results of Jansen et al. (1998). Consequently, a mathematical model cannot be defined and so an empirical distribution is used. USim was designed to operate with any empirical search string distribution and in our experiments the distribution obtained by Cacheda and Viña (21) was used, consisting of 26,654 search strings with their respective frequencies. For the rest of the search parameters (e.g. number of results displayed per page) the default values are used. The previous descriptive works by Jansen et al. (1998) and Cacheda and Viña (21) conclude that the high majority of the users do not change the default values in any search parameter. In case of browsing a category or visiting a document only an identifier is needed (of a category or a document, respectively). The best way to obtain a list of categories or documents identifiers is through the own simulation process, because there is no mathematical distribution that fits the categories or documents visited by the users (Cacheda and Viña, 21). Thus, the simulation tool obtains the categories and documents identifiers by either the searching or browsing processes, and then they are stored in their respective caches. Each identifier has a finite life in the cache, which is a configuration parameter (typically the average length of a user s session). When a request of a category or document is performed, the corresponding identifier is randomly selected from the cache. Once the request is generated it will be sent to the search engine using the HTTP protocol. The HTTPClient API (Tschalar, 23) for Java was used for this purpose. The URL for each type of request and the name of the parameters required must be defined in the configuration of USim. When the HTTP response has been received a simple and configurable parsing process is used to extract some relevant information from the HTML document received, but only for the searches or categories requests. As it was previously described, the categories and document identifiers are extracted from this response to be later stored in their respective simulation caches. And also two values are obtained from the result page: the number of categories and the total number of documents retrieved in the answer. Finally, for each request sent to the IR system the following information is stored in an output file for a subsequent analysis: - Timestamp: date and time when the answer was received from the retrieval system. - Request identifier: depending on the type of request, the search string or the category or document identifier. - Response time: the time since the request was sent until the response was received (the HTML document was completely received). - Images: the number of images included in the answer. - Response time images: the additional time needed to download the images. - And for the searches and categories requests, the number of categories and documents retrieved, obtained from the parsing process.

Figure 1: General configuration for USim Figure 2: Searches configuration for USim 3.2 Operation In this section we briefly describe the main functionality characteristics of this simulation tool, with the objective of speed up the understanding and comprehension of the performance evaluation situations described in the next section. Firstly it is important to mention that USim can operate with a user interface or in batch mode, using proprietary configuration files. The Figure 1 shows the graphic interface used to configure the general parameters that are related to the whole simulation process. In this part, the user can determine the length of the simulation and the life in cache of the category and document identificatiers gathered by USim during the simulation process. The configuration can be stored in order to use the simulation tool in the batch mode. When the simulation starts the application checks which types of requests must be sent to the IR system, and each module is started independently. In this way, USim can be used with the main types of IR system: Web directories, search engines and metasearch engines. If the system analyzed is a Web directory, USim must send searches and accesses to categories, meanwhile if the system analyzed is a search engine or a metasearch engine only searches must be sent. The module of visits to documents is included because some IR systems include a middle page between the search results and the final document, which also increases the load of the system. The remaining of the user interface is used to configure the parameters of each type of request. These parameters are very similar for the three types of requests and so we only describe the search configuration (see Figure 2). The main parameters are the number of searches per minute and the URL of the search system. Also, the names of the parameters needed to perform the search, commonly the search string, and the number of results to obtain and the position of the first result. The simulation tool will generate the values associated to these parameters and invoke the search engine passing these parameters and the generated values, using either a GET or POST method. The searches file contains the empirical distribution of the search strings that will be used by USim and, the output file will store all the information for the search requests (timestamp, request identifier, response time and so on). It is important to mention that the number of searches per minute is not static but it can dynamically change during the simulation using the parameters Increase in and every minute. This will help in the evaluation of the Web IR systems under different workloads using only one simulation process. The configuration of categories and documents is quite similar. In this case, the number of categories browsed and documents visited per minute are automatically estimated using the linear relationship obtained in (Cacheda and Viña, 21), although these and the rest of the parameters can be directly modified by the user. 4. PERFORMANCE EVALUATION RESULTS USim is a simulation tool designed and implemented for the performance evaluation of any type of Web IR system, especially for research purposes in a local environment with negligible network latency.

USim can be used and configured in two different ways to measure two different parameters of the performance of an IR system. The first one is the classical denial of service method that will measure the maximum number of requests supported by the system (named saturation threshold). The second one is more interesting for research purposes because it will measure the response time of a search engine under different workload situations. This leads directly to the comparison of the performance for variations of the same search engine or even completely different IR systems. 4.1 Saturation threshold One of the critical measures for any Web IR system is the maximum number of requests supported in a minute. It is evident that starting from a certain threshold the performance of the system decreases suddenly, increasing the response times up to the denial of service. This point is named saturation threshold. To establish the saturation threshold is fundamental to take some preventive actions (such as, application management techniques or the incorporation of new hardware) before the denial of service is produced. In this case, USim can be easily configured to simulate the effect of multiple simultaneous users sending requests in a controlled environment to the IR system. For this purpose, an experiment was designed where the saturation threshold must be measured for a prototype of a basic Web directory installed in a Sun Ultra Enterprise 25, with one 3 MHz CPU and 768 MB of main memory. This basic Web directory consists of approximately 1 categories and more than 5, classified documents, and its architecture is described in (Cacheda and Viña, 23). USim is configured to send requests to the IR system and to periodically increase the workload. For all the requests the response time is measured. The simulation tool is configured to start with 5 searches per minute (and 4.1 browsed categories per minute and 7.5 viewed documents per minute). The initial value of searches per minute is increased every 1 minutes in 1 search per minute (and also the equivalent increase is performed in the number of categories and documents). The results are showed in Figure 3 and Figure 4. The first graph (Figure 3) presents the response times of the searches performed to the system through the whole simulation. The image is quite clear: approximately above 21 searches per minute the response times start augmenting rapidly. At this point, every new query that is requested to the system will increase the load of the system and worsen the situation. This condition will stop when the number of requests per minute decreases. Figure 4 illustrates the number of errors pages returned by the system and obviously, the IR system operates perfectly until the number of searches is superior to 21 requests per minute. The simulation process also obtains the response times for the browsed categories and the viewed documents, but these graphs are quite similar and don t contribute with any new relevant information. Therefore, using USim the saturation threshold is established in 21 searches per minute (and the respective values for the categories browsed and viewed documents). 4 7 6 3 5 2 4 Response time (ms) 1 5. 1. 12. 14. 16. 18. 2. 21. 23. 24. 8. 11. 13. 15. 17. 19. 21. 22. 24. 25. Errors 3 2 1 5. 7. 9. 11. 13. 15. 17. 19. 21. 23. 25. Searches per minute Figure 3: Response time vs. searches per minute Searches per minute Figure 4: System errors vs. searches per minute

4.2 Performance comparison The main goal of the performance evaluation of an IR system is the measure of the response times to compare and analyze the effect of some changes in the search system, in order to distinguish the real improvements obtained. The Web IR systems are under different workload levels through the time, starting with periods of low workload until situations of high workload or even saturation. Obviously, the performance of the system depends on the workload of each moment. Therefore, the performance evaluation must be done considering different workload situations in order to elaborate a more complete and representative study. In the work described in (Cacheda and Viña, 23) the performance of three different indexing techniques must be compared. In this case, the performance of a type of searches characteristic of the Web directories, named restricted searches was analyzed. A restricted search is a common search, but the results must belong to one category or any of its descendants. The first indexing technique uses a basic architecture based on inverted files and constitutes our baseline. The other two indexing techniques (named, hybrid model with total information and hybrid model with partial information) are based on a hybrid model of signature files embedded into inverted files. The hybrid model with total information will use the hybrid data structure for all the categories of the Web directory, increasing 1% the size of the index. Whereas the hybrid model with partial information only applies the hybrid data structure to the categories of the first levels of the Web directory, reducing approximately in 5% the size of the index. For more details about the hybrid data model and the two variants defined refer to (Cacheda and Viña, 23) and (Cacheda and Baeza-Yates, 24). Each of these search algorithms has been developed and tested on a Sun Ultra Enterprise 25, with one 3 MHz CPU and 768 MB of main memory. The methodology for the performance comparison is based on two units of USim, which are executed simultaneously. The first one will generate the workload on the IR system and the second one will send the restricted queries to test the performance of the system. The performance evaluation is performed over five different workload situations: null, low, medium, high and saturation. The first unit of USim was configured to generate these static workloads with the following average values: - Null: searches/minute, browsed categories/minute and viewed documents/minute. - Low: 5 searches/minute, 4.1 browsed categories/minute and 7.5 viewed documents/minute. - Medium: 12 searches/minute, 8.5 browsed categories/minute and 16.2 viewed documents/minute. - High: 19 searches/minute, 12.9 browsed categories/minute and 24.9 viewed documents/minute. - Saturation: 23 searches/minute, 15.4 browsed categories/minute and 29.9 viewed documents/minute. For each workload, the first unit of USim will send requests to the IR system, and after a stabilization period the second unit of Usim will be executed to measure the response time of the restricted queries. This second unit will use a reduced set of queries specifically designed to analyze the effects of some relevant parameters on the response time. In this case, two parameters are considered: the number of results obtained by the query and the number of documents associated with the restricted category. Therefore, a set of eight queries retrieving from to 2 documents was selected and three different categories were selected to 16 2 14 12 16 1 12 Response time (msegs) 8 6 4 2 5 15 3 5 75 1 2 Basic Hybrid total info Hybrid parcial info Response time (msegs) 8 4 5 15 3 5 75 1 2 Basic Hybrid total info Hybrid parcial info Number of results Figure 5: Response time (null workload) Number of results Figure 6: Response time (low workload)

3 3 25 25 2 2 Response time (msegs) 15 1 5 5 15 3 5 75 1 2 Basic Hybrid total info Hybrid parcial info Response time (msegs) 15 1 5 5 15 3 5 75 1 2 Basic Hybrid total info Hybrid parcial info Number of results Number of results Figure 7: Response (medium workload) Figure 8: Response time (high workload) restrict the queries (with 2, 1 and 5 documents associated, respectively). In Figure 5, Figure 6, Figure 7 and Figure 8 we show the results obtained for the comparison of the three algorithms. All the experiments were analyzed using the ANOVA test. Three factors are defined in the ANOVA: the type of model (basic, hybrid model with total information and hybrid model with partial information), the number of results retrieved by the query and the number of documents associated with the restricted category. Obviously, the number of results and the number of documents associated with the restricted category are relevant factors of the response time, and the objective is to determine if the type of model is also relevant, comparing the performance of the hybrid models versus the basic model. Figure 5 represents an ideal situation, and it is clear that the hybrid models perform much better than the baseline. In fact, if the query gets more than 5 results, the response times are reduced in 5% in the hybrid models versus the basic model. The ANOVA test considered relevant the three factors analyzed (R square =.988). The same situation is represented in Figure 6 and Figure 7, but with low and medium workload, respectively. The behavior of the three algorithms is equivalent to the previous one, except that the response times are slightly higher (the three factor are also relevant in the ANOVA test, with R square =.923 and R square =.92, respectively). But the situation changes in Figure 8 where a high workload is generated in the IR system. The ANOVA test still considers relevant the three factors, with a high R square (R square =.857). The most relevant aspect is that the performance of the hybrid model with total information is deteriorated, performing similarly to the baseline. While, the hybrid model with partial information keeps the performance improvement on 5% versus the baseline, and now also versus the hybrid model with total information. The hybrid data structure defined on both hybrid models seems to clearly improve the performance for the restricted queries. Although, in a high workload environment the disk operations are the bottleneck and so, the hybrid model with total information, with its higher index size, is penalized. The saturation workload is not described because its ANOVA test shows that only a small part of the variation in the response times is explained (R square =.536) and therefore its results are not significant. This experiment demonstrates the importance of considering the workload in any IR system, and specifically in the Web IR systems. Initially, both hybrid models behaved in a similar way, performing 5% better than the baseline. But, in the end, we have found out that only the hybrid model with partial information is able to keep the improvement in the performance in all the circumstances, whereas the hybrid model with total information decreases its performance in high workload situations, due to its higher disk requirements. 5. CONCLUSION With the emergence of the Web IR systems some new measures were defined for the evaluation of the retrieval (relative precision, relevant/useful pages, etc.). But this must also involve the performance evaluation of these retrieval systems. Therefore, in this paper we have presented USim, a simulation tool for the performance evaluation of Web IR systems based on the simulation of users behavior. This simulation

tool helps in the performance evaluation of Web IR systems in two different ways: estimating the saturation threshold of the system and comparing the performance of different search algorithms or engines. To establish the saturation threshold is fundamental before any Internet search engine is put into production because it will estimate the maximum load that the system will be able to bear before degrading its performance. So, before this point is reached, some preventive actions can be taken to increase the processing capacity of the system or to avoid this threshold using application management techniques. The second point is the most interesting because, traditionally, to measure the response time of any IR system has been fundamental in the performance evaluation. However, the Internet search engines must operate under different workload situations, with a high number of concurrent users. In this way, the response times must be measured considering different workload situations; otherwise erroneous conclusions can be achieved. The results obtained prove that a comparison only in null workload environments is not enough because we have discovered how the performance of a search algorithm seemed to be appropriate in a low workload environment, whereas its performance decreases suddenly in a high workload situation. From a general point of view, with USim we intend to create an approximation to some new performance evaluation techniques specifically developed for the Internet search engines. The use of the simulation for the performance evaluation of Internet search engines seems promising; mainly because the response times can be estimated more accurately considering different workload environments. For further research, an interesting point is the extension of USim to operate on a WAN. At the moment, the main limitation of our simulation tool is to operate on a LAN, where the network latency is negligible. However, on a WAN the network latency must be estimated to obtain the actual search engine response time. Also, in our future work we will concentrate in the improvement of USim in order to make it publicly available, trying to develop a more generic application suitable to any type of retrieval system. This implies that an advance must be done in the information extraction of the results Web pages (where the use of XML seems promising) and in the different parameters used by the search engines URLs. REFERENCES Cacheda, F. and Viña, A., 21. Experiencies retrieving information in the World Wide Web. Proceedings of the 6th IEEE Symposium on Computers and Communications. Hammamet, Tunisia, pp. 72-79. Cacheda, F. and Viña, A., 23, Optimization of Restricted Searches in Web Directories Using Hybrid Data Structures. In Lecture Notes on Computer Science Vol. 2633, pp. 436-451. Cacheda, F. and Baeza-Yates, R., 24, An Optimistic Model for Searching Web Directories. In Lecture Notes on Computer Science Vol. 2997, pp. 394-49. Hawking, D. et al., 1999. Results and challenges in Web search evaluation. In Proceedings of the 8th World Wide Web Conference, Toronto, Canada, pp. 243-252. Jansen, B. and Pooch, U., 21. Web User Studies: A Review and Framekwork for Future Work. In Journal of the American Society of Information Science and Technology, Vol. 52, No 3, pp. 235-246. Jansen, B. et al., 1998. Real Life Information Retrieval: A Study Of User Queries On The Web. In SIGIR FORUM, Vol. 32, No 1, pp. 5-17. Kirsch, S., 1998. Infoseek s experiences searching the Internet. In SIGIR FORUM, Vol. 32, No. 2, pp. 3-7. TREC, 24. Text REtrieval Conference, NIST, TREC home page. http://trec.nist.gov/. Ozmultu, H.C., Spink, A. and Ozmultu, S., 22. Analysis of large data logs: an application of Poisson sampling on Excite web queries. In Information Processing and Management Vol. 38, pp. 473-49. Silverstein, C. et al., 1999. Analysis of a Very Large Web Search Engine Query Log. In SIGIR FORUM, Vol. 33, No. 1, pp. 6-12. Spink, A. and Ozmultu, H. C., 22. Characteristics of question format web queries: an exploratory study. In Information Processing and Management Vol. 38, pp. 453-471. A. Spink, B. Jansen, D. Wolfram, T. Saracevic, From E-sex to E-commerce: Web Search Changes. IEEE Computer, 35(3): 17-111, 22. R. Tschalar, HTTPClient. http://www.innovation.ch/java/httpclient/, 23. Zipf, G., 1949. Human behaviour and the principle of least effort. Ed. Addison-Wesley. Zobel, J., Moffat, A. and Ramamohanarao, K., 1996. Guidelines for Presentation and Comparison of Indexing Techniques. In ACM SIGMOD Record, Vol. 25, No. 3, pp. 1-15.