Mercure at trec6 2 IRIT/SIG. Campus Univ. Toulouse III. F Toulouse. fbougha,

Similar documents
Evaluating a Conceptual Indexing Method by Utilizing WordNet

Probabilistic Learning Approaches for Indexing and Retrieval with the. TREC-2 Collection

AT&T at TREC-6. Amit Singhal. AT&T Labs{Research. Abstract

Routing and Ad-hoc Retrieval with the. Nikolaus Walczuch, Norbert Fuhr, Michael Pollmann, Birgit Sievers. University of Dortmund, Germany.

Information Retrieval Research

Siemens TREC-4 Report: Further Experiments with Database. Merging. Ellen M. Voorhees. Siemens Corporate Research, Inc.

TREC-3 Ad Hoc Retrieval and Routing. Experiments using the WIN System. Paul Thompson. Howard Turtle. Bokyung Yang. James Flood

TREC-7 Experiments at the University of Maryland Douglas W. Oard Digital Library Research Group College of Library and Information Services University

CLEF-IP 2009: Exploring Standard IR Techniques on Patent Retrieval

TREC 2016 Dynamic Domain Track: Exploiting Passage Representation for Retrieval and Relevance Feedback

Term Frequency Normalisation Tuning for BM25 and DFR Models

TREC-10 Web Track Experiments at MSRA

Amit Singhal, Chris Buckley, Mandar Mitra. Department of Computer Science, Cornell University, Ithaca, NY 14853

A RECOMMENDER SYSTEM FOR SOCIAL BOOK SEARCH

indexing and query processing. The inverted le was constructed for the retrieval target collection which contains full texts of two years' Japanese pa

Performance Analysis of Information Retrieval Systems

nding that simple gloss (i.e., word-by-word) translations allowed users to outperform a Naive Bayes classier [3]. In the other study, Ogden et al., ev

An Evaluation of Information Retrieval Accuracy. with Simulated OCR Output. K. Taghva z, and J. Borsack z. University of Massachusetts, Amherst

Chinese track City took part in the Chinese track for the rst time. Two runs were submitted, one based on character searching and the other on words o

A Balanced Term-Weighting Scheme for Effective Document Matching. Technical Report

Rank-Ordering Documents According to Their Relevance in Information Retrieval Using Refinements of Ordered-Weighted Aggregations

Verbose Query Reduction by Learning to Rank for Social Book Search Track

where w t is the relevance weight assigned to a document due to query term t, q t is the weight attached to the term by the query, tf d is the number

Query Expansion with the Minimum User Feedback by Transductive Learning

The two successes have been in query expansion and in routing term selection. The modied term-weighting functions and passage retrieval have had small

M erg in g C lassifiers for Im p ro v ed In fo rm a tio n R e triev a l

RMIT University at TREC 2006: Terabyte Track

30000 Documents

CS54701: Information Retrieval

where NX qtf i NX = 37:4 ql :330 log dtf NX i dl + 80? 0:1937 log ctf i cf (2) N is the number of terms common to both query and document, qtf

This is an author-deposited version published in : Eprints ID : 12965

Robust Relevance-Based Language Models

James P. Callan and W. Bruce Croft. seven elds describing aspects of the information need: the information need that is related to, but often distinct

CS473: Course Review CS-473. Luo Si Department of Computer Science Purdue University

A Practical Passage-based Approach for Chinese Document Retrieval

Pseudo-Relevance Feedback and Title Re-Ranking for Chinese Information Retrieval

IITH at CLEF 2017: Finding Relevant Tweets for Cultural Events

Chapter 6: Information Retrieval and Web Search. An introduction

Real-time Query Expansion in Relevance Models

Report on the TREC-4 Experiment: Combining Probabilistic and Vector-Space Schemes

ResPubliQA 2010

A Voting Method for XML Retrieval

Web Search Engine Question Answering

Fondazione Ugo Bordoni at TREC 2004

An Attempt to Identify Weakest and Strongest Queries

Comparison of Two Interactive Search Refinement Techniques

Applying the KISS Principle for the CLEF- IP 2010 Prior Art Candidate Patent Search Task

Prior Art Retrieval Using Various Patent Document Fields Contents

Performance Measures for Multi-Graded Relevance

James Mayfield! The Johns Hopkins University Applied Physics Laboratory The Human Language Technology Center of Excellence!

Reducing Redundancy with Anchor Text and Spam Priors

Report on TREC-9 Ellen M. Voorhees National Institute of Standards and Technology 1 Introduction The ninth Text REtrieval Conf

Information Retrieval. Techniques for Relevance Feedback

Informativeness for Adhoc IR Evaluation:

characteristic on several topics. Part of the reason is the free publication and multiplication of the Web such that replicated pages are repeated in

IIT at TREC-10. A. Chowdhury AOL Inc. D. Holmes NCR Corporation

A Prototype for Integrating Probabilistic Fact. and Text Retrieval

Component ranking and Automatic Query Refinement for XML Retrieval

Ranking Function Optimizaton Based on OKAPI and K-Means

Inter and Intra-Document Contexts Applied in Polyrepresentation

Interface. Dispatcher. Meta Searcher. Index DataBase. Parser & Indexer. Ranker

This is an author-deposited version published in : Eprints ID : 15246

CLARIT Compound Queries and Constraint-Controlled Feedback in TREC-5 Ad-Hoc Experiments

523, IEEE Expert, England, Gaithersburg, , 1989, pp in Digital Libraries (ADL'99), Baltimore, 1998.

From Passages into Elements in XML Retrieval

The University of Amsterdam at the CLEF 2008 Domain Specific Track

Risk Minimization and Language Modeling in Text Retrieval Thesis Summary

CS47300: Web Information Search and Management

Information Retrieval Term Project : Incremental Indexing Searching Engine

VIDEO SEARCHING AND BROWSING USING VIEWFINDER

A Methodology for End-to-End Evaluation of Arabic Document Image Processing Software

University of Waterloo: Logistic Regression and Reciprocal Rank Fusion at the Microblog Track

Hummingbird's Fulcrum SearchServer at CLEF 2001

AT&T at TREC-7. Amit Singhal John Choi Donald Hindle David D. Lewis. Fernando Pereira. AT&T Labs{Research

Table 1: Organizations participating in TREC-8 ACSys AT&T Labs Research CL Research CLARITECH Corporation Cambridge University Carnegie Mellon Univers

Federated Text Search

Making Retrieval Faster Through Document Clustering

Melbourne University at the 2006 Terabyte Track

A New Approach for Automatic Thesaurus Construction and Query Expansion for Document Retrieval

DATABASE MERGING STRATEGY BASED ON LOGISTIC REGRESSION

Content-Based Image Retrieval By Relevance. Feedback? Nanjing University of Science and Technology,

University of Glasgow at TREC2004: Experiments in Web, Robust and Terabyte tracks with Terrier

Finding Relevant Documents using Top Ranking Sentences: An Evaluation of Two Alternative Schemes

Fondazione Ugo Bordoni at TREC 2003: robust and web track

CS54701: Information Retrieval

Document Expansion for Text-based Image Retrieval at CLEF 2009

Overview of FIRE 2011 Prasenjit Majumder on behalf of the FIRE team

Better Contextual Suggestions in ClueWeb12 Using Domain Knowledge Inferred from The Open Web

Web Information Retrieval. Exercises Evaluation in information retrieval

Building Test Collections. Donna Harman National Institute of Standards and Technology

Block Addressing Indices for Approximate Text Retrieval. University of Chile. Blanco Encalada Santiago - Chile.

Document Filtering Method Using Non-Relevant Information Profile

Information Retrieval: Retrieval Models

Estimating Embedding Vectors for Queries

highest cosine coecient [5] are returned. Notice that a query can hit documents without having common terms because the k indexing dimensions indicate

Information Retrieval

Maximal Termsets as a Query Structuring Mechanism

TEXT CHAPTER 5. W. Bruce Croft BACKGROUND

Favorites-Based Search Result Ordering

X. A Relevance Feedback System Based on Document Transformations. S. R. Friedman, J. A. Maceyak, and S. F. Weiss

Transcription:

Mercure at trec6 M. Boughanem 1 2 C. Soule-Dupuy 2 3 1 MSI Universite de Limoges 123, Av. Albert Thomas F-87060 Limoges 2 IRIT/SIG Campus Univ. Toulouse III 118, Route de Narbonne F-31062 Toulouse 3 CERISS Universite Toulouse I Manufacture des Tabacs F-31000 Toulouse Email : fbougha, souleg@irit.fr 1 Introduction We continue our work in trec performing runs in adhoc, routing and part of the cross language track. The major investigations this year are the weight schemes modication to take into account the document length. We also experiment the high precision procedure in automatic adhoc environment by tuning the term weight parameters. 2 Mercure model Mercure is an information retrieval system based on a connexionist approach and modelled by a network (as shown in the gure 1) containing an input representing the query, a term layer representing the indexing terms, a document layer representing the documents and an output representing the retrieved documents. The term nodes (or neurons) are connected to the document nodes (or neurons) by weighted indexing links. Mercure includes the implementation of two main components : the query evaluation based on spreading activation from the input to the output through the indexing links and the automatic query modication based on backpropagation of the document relevance. 2.1 Query evaluation based on spreading activation The query evaluation is performed as follows : 1. Build the input Input k = (q 1k ; q 2k ; :::; q Tk ), 2. Apply this input to the term layer. Each term neuron computes an input value : In(N ti ) = q ik and then an output value : Out(N ti ) = g(in(n ti )) 3. These signals are propagated forwards through the network. Each neuron computes an input and an output value : then, Out(N Di ) = g(in(n Di )) In(N Di ) = TX j=1 Out(N j ) w ij 1

Information need Query? (t1, t2) Input layer q1j, q2j,, q3i, qtj New input Nt1 c12 Nt2 Nti NtT Propagation process for query evaluation Neuron term layer w11 wi1 Backpropagation of the relevance for query modification Neuron document layer ND1 ND2 ND3 NDj NDM Output layer Desired output D1 D2 D3 D5 Retrieved documents Judged documents Figure 1: The Mercure Model. The output vector is : Output k (Out(N D1 ); Out(N D2 ); :::; Out(N DM )) These output values computed by the document neurons are used to rank the list of retrieved documents. 2.2 Query modication based on relevance backpropagation The automatic query modication is based on spreading the document relevance values backwards the network. The retrieved documents are used to build the DesiredOutput. To each judged document is assigned a relevance value. A positive relevance value is assigned to relevant documents, a negative value to non-relevant documents. The desired output is represented by the vector of the form : DesiredOutput = (rel 1 ; :::; rel i ; :::; rel M ). This strategy consists in backpropagating the relevance values from the output layer to the input layer, and it is performed as follows : 1. Build the desired output : DesiredOutput = (rel 1 ; :::; rel i ; :::; rel M ), 2. Apply this output to the neuron document layer. Each neuron computes an input value : In(N Di ) = rel i and then an ouput signal : Out(N Di ) = g(in(n Di )) 3. The output signals are backpropagated to the term neuron layer. Each neuron term computes an input value : In(N ti ) = MX j=1 and then an output signal : Out(N ti ) = g(in(n ti )) (w ij Out(N Dj )) 4. A new input is then computed according to this formula : N ewinput k = Input k + Out(N t ) 2

This new input is applied to the term neuron layer and a new query evaluation is then done. Several formulations can be used to construct the desired output. For this experimentation we have chosen the following formula : - for relevant document : rel i = Coef Rel Nb rel - for nonrelevant document : rel i = Coef NRel Nb Nrel Where : Coef Rel, Coef N Rel : relevance coecient of the documents (positive for relevant and negative for non-relevant documents), N b rel, N b N rel : number of relevant and non-relevant documents respectively, 3 General Investigations Our rst investigation is to modify the indexing weight to take into account the document length. Our formula is inspired by Okapi and Smart term weight functions. It is expressed by : w ij = (1+log(tf ij )) 1+log(average j (tfij)) (h 1 + h 2 log( N n i )) h 3 + h 4 The query term weight in the input is expressed by : q ik = doclenj avg doclen (1 + log(tf ik )) (log(n=n i ) q PT j=1 (1 + log(tf jk)) (log(n=n j ) 2 Where : w ij : the weight of the link between the term t i and the document D j, tf ij : the frequency of the term t i in the document D j, T : the number of documents in the collection, n i : the number of documents containing the term t i, doclen j : document length in words (without stop words), avg doclen : average document length, computed for each database. 4 Adhoc experiment and results 4.1 adhoc methodology Our investigation is to improve the query expansion in automatic adhoc environment. The "blind" relevance feedback was performed by assuming the top retrieved documents as relevant and the low retrieved as non relevant. Some eorts have been undertaken to improve the precision in the small top ranked documents. The basic goal is to produce the "High precision" by "trading" the recall for the precision, [4] [5] (e.g we can loose some relevant documents if we are sure that the remaining ones are relevant). A way to produce a high precision could be by using "good" query term and document term weights. Our strategy in adhoc trec-6 is to weight the indexing links in order to maximize the precision at small ranked top documents and then a "normal" weight scheme (weight performing a best precision at 1000 top ranked documents) will be used in the relevance backpropagation process and in the new input spreading. The weight schemes we used in trec-6 are obtained by tuning the h 1, h 2, h 3, h 4 parameters. 3

Series of experiments have been undertaken on TREC-5 database and queries. The parameters we have chosen to use in TREC-6 experiment are : h 1 = 1, h 2 = 0, h 3 = :8, h 4 = :2 for the high precision and h 1 = :8, h 2 = :2, h 3 = :8, h 4 = :2 for what we called a "normal" weight. The remaining parameters used in the relevance backpropagation are : Coef Rel = 1, Coef N Rel =?:75, = 2, = :5, N b rel = 12, N b N rel = 500 (from 501 to 1000). 4.2 Adhoc results and discussion Preliminary investigations The rst result we underline concerns the term weight functions. The table 1 shows the average precision of basic run obtained by some IR systems in TREC-5. We can notice that the weight schemes we used are quite good (h 1 = :8, h 2 = :2, h 3 = :8, h 4 = :2). TREC-5 results system average precision in initial search Mercure : 0.1578 Okapi : 0.1520 Smart : 0.1484 Inquery : 0.1442 Table 1: Comparative basic search trec-5 results Automatic adhoc results Three automatic runs were submitted : Mercure2 (description only), Mercure1 (long topic : title, description and narrative) and Mercure3 (title only). These runs were based on completely automatic processing of TREC queries and automatic query expansion, the high precision concept was also used. Table 2 compares our runs against the published median runs. We notice that most of the runs are above the median. TREC results Run Best median < median Mercure2 (description) 1 40 10 Mercure3 (title) 1 29 18 Mercure1 (long topic) 5 44 6 Table 2: Comparative automatic adhoc results at average precision We unfortunatelly noticed an error in the script that has been used to perform the adhoc description run (the other runs are right). The weight scheme (i.e. the h i parameters) used to produce the high precision has also been used by mistake in the relevance backpropagation process instead of the "normal" h i values. The table 3 shows the ocial and the corrected runs). We actually notice a dierence between the description runs, the other runs seem good. The table 4 the average precisions of the basic run using the high precision and the run after query expansion on the three corrected runs. The query expansion is done by using the following values of Mercure parameters : N b rel = 12, N b N rel = 500, 501-1000 non-relevant documents and the number 4

Ocial results corrected results Run Average precision R. Precision Average precision R. Precision Mercure2 (description) 0.1640 0.2065 0.1720 0.2108 Mercure3 (title) 0.2316 0.2689 0.2316 0.2689 Mercure1 (long topic) 0.2305 0.2700 0.2305 0.2700 Table 3: Automatic adhoc results - 50 queries pf term added to the query is 16. We notice that the automatic query expansion is still eective in the adhoc environment. Run average precision Mercure3 : title only basic search using h i producing the high precision 0.2041 Exp. N b rel = 12, N b N rel = 500, 501-1000 non-relev docs 0.2316 (+13.47 %) (Mercure2.C) description only basic search using h i producing the high precision 0.1549 Exp. N b rel = 12, N b N rel = 500, 501-1000 non-relev docs 0.1710 (+10.39 %) (Mercure1) long topic basic search using h i producing the high precision 0.2128 Exp. N b rel = 12, N b N rel = 500, 501-1000 non-relev docs 0.2305 (+8.32 %) Table 4: Adhoc component results - 50 queries However we notice that the way used to improve the precision at top ranked documents did not have a positive eect as in the trec-5 adhoc. Indeed, the table 5 shows the results in the description run (Mercure2.C.N) when using the "normal" h i values. We observe a slight dierent in favour of the Mercure2.C.N run. We do not yet analyze the results of the title and long topics runs. Run average precision Mercure2.C.N: description only basic search using "normal" h i 0.1693 Exp. N b rel = 12, N b N rel = 500, 501-100 non-rel docs 0.1772 Table 5: Adhoc component results - 50 queries 5

5 Routing experiment and results All trec-6 training data were used (relevant and non relevant documents). The queries are initially built automatically from all the elds of the topics and then expanded by using the 30 top terms resulting from the relevance backpropagation procedure. Each query was evaluated by varying the dierent Mercure parameters, h i and,, etc. The queries performing the best average precision in the training data were selected. Moreover, a slight modication has been performed in the relevance value formula, it concerns the positive relevance value. Indeed, we decided to take into account the fact that a relevant document is or not among the 1000 retrieved documents in the initial search. The relevance value assigned to each relevant document becomes : rel i = coef R Nb rel BOOT BOOT = 1 if the relevant document is not in the 1000 documents BOOT < 1 if relevant document is retrieved (BOOT = :9 for routing trec6) no modication if a document is nonrelevant As the retrieved relevant documents are already close to the initial query, we give to the terms occurring in the non retrieved relevant documents more eect in the nal query building. The table 6 compares our routing runs against the medians published runs, more than 60% of queries are above the median. TREC routing results Run Best median < median Mercure4 1 29 18 Table 6: Comparative TREC Results at average precision The table 7 shows the dierence between the run based on the initial queries and the one based on the routing queries. We ha have no time to analyze these results TREC routing results Run average precision R precision Total Rel retrieved Mercure4 0.3061 0.3400 4774 Table 7: Comparative TREC Results at average precision Run average precision basic search (with the initial queries) 0.2676 Ocial run 0.3061 Table 8: Routing component results 47-queries 6

6 Cross language track : french to french Two runs french to french were submitted in CLIR track. The indexing and search methodologies are the same than the adhoc trec6 except the stemming algorithm where a cuto stemming method (7 characters) has been used. This stemming method has been implemented in all of our operational information retrieval systems dealing with french documents and french queries. The results obtained untill now lead us to go on the experiments with this stemming method. Moreover, for the adhoc task the high precision procedure has not been used because there is no relevance information to tune the weight scheme. The same parameters were used for the indexing weight h 1 = :8, h 2 = :2, h 3 = :8, h 4 = :2. The table 9 compares our runs against the published median runs. Most of the queries are above the median. TREC-6 cross language french to french Run Best median < median MercureFFs (description) 0 18 3 MercureFFl (long topic) 4 17 4 Table 9: Comparative TREC cross language at average precision The table 10 shows that the average precision and the R-precision for the dierent runs are quite good. Run Average precision R. Precision Total Rel Retrieved MercureFFs (description) 0.3619 0.3848 1023 MercureFFl (long topic) 0.3778 0.4015 1033 Table 10: cross language (french to french) results - 21 queries The important point we discuss concerns the automatic query expansion. Indeed, the table 11 shows the improvment obtained between the basic run and the run with an automatic query expansion using the following values of Mercure parameters : N b rel = 15, N b N rel = 500, 501-1000 non-relevant docs and the number of added terms is 16. In both, MercureFFs and MercureFFl the improvement about 10%. Run average precision description only basic search 0.3262 Expansion N b rel = 15, N b N rel = 500, 501-1000 non-relev docs 0.3619 (11%) long topic basic search 0.3479 Expansion N b rel = 15, N b N rel = 500, 501-1000 non-relev docs 0.3778 ( 8.6 %) Table 11: Adhoc cross language component results - 21 queries 7

7 Conclusion Last year, we participated in trec-5 in the adhoc and routing tasks in category B. Our main eort this year has been to participate in trec-6 in category A. We performed completely automatic runs in adhoc, routing and a part of the cross language tasks. At rst we planed to try, the passage retrieval, the data mining techniques [7] and the genetic algorithms [1] to automatically expand the queries. But nally, our investigations were the improvement of the term weighting and the automatic query modication. We spent much time on these experiments and decided to difer the planed experiments until the next year. However, the results we obtained for the main tasks are still encouraging this year. Our participation to the CLIR track was limited to a french to french experimentation to train our french language processing. Our goal now is to go on with a real cross language experiment. References [1] L. Tamine Reformulation de requ^etes basee sur l'algorithmique genetique Proceedings of INFORSID'97 Toulouse Juin 1997. [2] M. Boughanem & C. Soule-Dupuy, Query modication based on relevance backpropagation, Proceedings of the 5th International Conference on computer-assisted information searching on Internet (RIAO'97), Montreal, June 1997. [3] M. Boughanem & C. Soule-Dupuy, Mercure : adhoc and routing tasks, 5th International Conference on Text REtrieval TREC2, Harman D.K. (Ed.), NIST SP 500-236, 1996. [4] C. Buckley & al, Query zoning : TREC'5, 5th International Conference on Text REtrieval TREC2, Harman D.K. (Ed.), NIST SP 500-236, 1996. [5] B. croft, & al, INQUERY at TREC-5. 5th International Conference on Text REtrieval TREC5, Harman D.K. (Ed.), 1996. [6] S. Robertson and al, Okapi at TREC-5. 5th International Conference on Text REtrieval TREC2, Harman D.K. (Ed.), NIST SP 500-236, 1996. [7] T. Dkaki, B. Dousset & M. Mothe, Mining information in order to extract hidden and strategical information, Proceedings of the 5th International Conference on computer-assisted information searching on Internet (RIAO'97), Montreal, June 1997. 8