Automatic Search Engine Evaluation with Click-through Data Analysis. Yiqun Liu State Key Lab of Intelligent Tech. & Sys Jun.

Similar documents
Automatic Query Type Identification Based on Click Through Information

Automatic Query Type Identification Based on Click Through Information. State Key Lab of Intelligent Tech. & Sys Tsinghua University

Identifying Web Spam With User Behavior Analysis

HYBRIDIZED MODEL FOR EFFICIENT MATCHING AND DATA PREDICTION IN INFORMATION RETRIEVAL

Information Retrieval Spring Web retrieval

Automatic Identification of User Goals in Web Search [WWW 05]

Data Cleansing for Web Information Retrieval Using Query Independent Features

Modern Retrieval Evaluations. Hongning Wang

Estimating Credibility of User Clicks with Mouse Movement and Eye-tracking Information

TREC OpenSearch Planning Session

Welcome to the class of Web Information Retrieval!

CHAPTER THREE INFORMATION RETRIEVAL SYSTEM

Ph.D. in Computer Science & Technology, Tsinghua University, Beijing, China, 2007

CSCI 599: Applications of Natural Language Processing Information Retrieval Evaluation"

Information Retrieval May 15. Web retrieval

Adaptive Search at Essex

Advanced Search Techniques for Large Scale Data Analytics Pavel Zezula and Jan Sedmidubsky Masaryk University

Constructing Websites toward High Ranking Using Search Engine Optimization SEO

Inferring User Search for Feedback Sessions

NTU Approaches to Subtopic Mining and Document Ranking at NTCIR-9 Intent Task

Information Retrieval

Building Test Collections. Donna Harman National Institute of Standards and Technology

CS473: Course Review CS-473. Luo Si Department of Computer Science Purdue University

In the recent past, the World Wide Web has been witnessing an. explosive growth. All the leading web search engines, namely, Google,

NUSIS at TREC 2011 Microblog Track: Refining Query Results with Hashtags

Outline. Structures for subject browsing. Subject browsing. Research issues. Renardus

Domain Specific Search Engine for Students

A Metric for Inferring User Search Goals in Search Engines

Lec 8: Adaptive Information Retrieval 2

Understanding the Query: THCIB and THUIS at NTCIR-10 Intent Task. Junjun Wang 2013/4/22

A web directory lists web sites by category and subcategory. Web directory entries are usually found and categorized by humans.

Evaluation of Retrieval Systems

IJREAT International Journal of Research in Engineering & Advanced Technology, Volume 1, Issue 5, Oct-Nov, ISSN:

Entity and Knowledge Base-oriented Information Retrieval

ICTNET at Web Track 2010 Diversity Task

Connecting Pure to the web

Northeastern University in TREC 2009 Million Query Track

SEO ISSUES FOUND ON YOUR SITE (JANUARY 2, 2017)

Retrieval Evaluation

Comparative Analysis of Clicks and Judgments for IR Evaluation

Information Retrieval II

How Do Sponsored Search Results Affect User Behavior in Web Search?

Monthly SEO Report. Example Client 16 November 2012 Scott Lawson. Date. Prepared by

Web Data mining-a Research area in Web usage mining

whitehvac.com SEO ISSUES FOUND ON YOUR SITE (JANUARY 2, 2017)

Search Engines Chapter 8 Evaluating Search Engines Felix Naumann

power up your business SEO (SEARCH ENGINE OPTIMISATION)

Company Snapshot. $ million $2.46 billion 39% 52% $418.5 million

Information Search in Web Archives

Expected Learning Objective Activities/FAs Planned

Relevant?!? Algoritmi per IR. Goal of a Search Engine. Prof. Paolo Ferragina, Algoritmi per "Information Retrieval" Web Search

Introduction to Information Retrieval. Hongning Wang

ISE Cyber Security UCITS Index (HUR)

Searching the Deep Web

Context based Re-ranking of Web Documents (CReWD)

Database Design and Management - BADM 352 Fall 2009 Syllabus and Schedule

Searching the Deep Web

Retrieval Evaluation. Hongning Wang

Web Search Basics. Berlin Chen Department t of Computer Science & Information Engineering National Taiwan Normal University

News Page Discovery Policy for Instant Crawlers

Search Marketing 101 CCT332

Information Retrieval

5 Choosing keywords Initially choosing keywords Frequent and rare keywords Evaluating the competition rates of search

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

Web Usage Mining: A Research Area in Web Mining

Overview of Web Mining Techniques and its Application towards Web

Informativeness for Adhoc IR Evaluation:

Information Retrieval

Query Disambiguation from Web Search Logs

China Internet Market and Online Audience Measurement

Information Retrieval. Lecture 7 - Evaluation in Information Retrieval. Introduction. Overview. Standard test collection. Wintersemester 2007

Applying the KISS Principle for the CLEF- IP 2010 Prior Art Candidate Patent Search Task

Information Retrieval

SEO and UAEX.EDU GETTING YOUR WEB PAGES FOUND IN GOOGLE

The CPA Exam and Requirements. Adapted and modified from material originally created by David Reinus.

Web Search Basics. Berlin Chen Department of Computer Science & Information Engineering National Taiwan Normal University

Examining the Limits of Crowdsourcing for Relevance Assessment

Evaluation of Retrieval Systems

IMPROVING INFORMATION RETRIEVAL BASED ON QUERY CLASSIFICATION ALGORITHM

DATA MINING II - 1DL460. Spring 2014"

DATA MINING - 1DL105, 1DL111

Identifying Web Spam with the Wisdom of the Crowds

SEO ISSUES FOUND ON YOUR SITE (DECEMBER 30, 2016)

Enhancing Cluster Quality by Using User Browsing Time

On Duplicate Results in a Search Session

University of Virginia Department of Computer Science. CS 4501: Information Retrieval Fall 2015

THE SCIENCE OF SEO A MANUAL FOR THE MODERN PR PROFESSIONAL

App Economy Market analysis for Economic Development

CS6200 Information Retrieval. Jesse Anderton College of Computer and Information Science Northeastern University

CS290N Summary Tao Yang

Chapter 4. Processing Text

Measurement and Tracking Awareness June 2011

Information Push Service of University Library in Network and Information Age

Evaluation of Retrieval Systems

Web Database Integration

SEO ISSUES FOUND ON YOUR SITE (JANUARY 2, 2017)

Media Pack. Home About User Benefits Advertiser Benefits Advertising Packages

Accessing XML documents: The INEX initiative. Mounia Lalmas, Thomas Rölleke, Zoltán Szlávik, Tassos Tombros (+ Duisburg-Essen)

NTCIR-13 Core Task: Short Text Conversation (STC-2)

Maximising Search Engine Marketing. Search Marketing and Google Analytics

Transcription:

Automatic Search Engine Evaluation with Click-through Data Analysis Yiqun Liu State Key Lab of Intelligent Tech. & Sys Jun. 3th, 2007

Recent work: Using query log and click-through data analysis to: identify search engine users information need types evaluate search engine performance automatically separate key resource pages from others estimate Web page quality Our Lab: A joint lab Yiqun Liu Ph.D student from Tsinghua University, Beijing, China. liuyiqun03@gmail.com R&D Support to a widely-used Chinese Search Engine Sogou.com, platform to get research results realized.

Yiqun Liu Ph.D student from Tsinghua University, Beijing, China. liuyiqun03@gmail.com Web Data Cleansing Using query-independent features and ML algorithms 5% web pages can meet >90% user s search needs Query type identification Identify the type of user s information need Over 80% queries are correctly classified Search engine performance evaluation Construct query topic set and answer set Automatically. Obtain similar evaluation results with manual based methods, and cost far less time and labor.

Introduction Lots of search engines offer services on the Web Search Engine Performance Evaluation Web Users over 120 million users in mainland Search Advertisers spending 5.6 billion RMBs in 2007 Search engineers and researchers

Introduction Evaluation is a key issue in IR research Evaluation became central to R&D in IR to such an extent that new designs and proposals and their evaluation became one. (Saracevic, 1995) Cranfield-like evaluation methodology Proposed by Cleverdon et al in 1966. A set of query topics, their corresponding answers (usually called qrels) and evaluation metrics. Adopted by IR workshops such as TREC and NTCIR.

Introduction Problems with Web IR evaluation 9 people months are required to judge one topic for a collection of 8 million documents. (Voorhees, 2001) Search engines (Yahoo!, Google) index over 10 billion Web documents. Almost Impossible to use human-assessed query and qrel sets in Web IR system evaluation.

Related works Efforts in automatic search engine performance evaluation (Cranfield-like) Considering pseudo feedback documents as correct answers (Soboroff, 2001; Nuray, 2003) Adopting query topics and qrels extracted from Web page directories such as open directory project (ODP) (Chowdhury, 2002; Beitzel, 2003)

Related works Efforts in automatic search engine performance evaluation (other evaluation approaches) Term Relevance Sets (Trels) method. Define a pre-specified list of terms relevant and irrelevant to these queries. (Amitay, 2004) The use of click-through data. Construct a unified meta search interface to collect users behaviour information. (Joachims, 2002)

Our method A cranfield-like approach Accepted by major IR research efforts Difficulty: annotating all correct answers automatically Click-through behavior analysis Single user may be cheated by search spams or SEOs. User group s behavior information is more reliable.

Automatic Evaluation Process Information need behind user queries Proposed by Broder (2003) Navigational type: One query have only one correct answer. Informational type: One query may have several correct answers. Different behavior over different types of information needs

Information needs and Evaluation Informational queries cannot be annotated People click different answers while using different search engines. 0.35 0.3 baidu google yahoo sogou 0.25 0.2 0.15 0.1 0.05 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27

Automatic Evaluation Process

Query Set Classification Less Effort Assumption & N Clicks Satisfied (ncs) Evidence 60% While performing a navigational type search request, Navigational Informational & Transactional 50% user tend to click a small number of URLs in the result list. 40% 30% 20% 10% 0% 0.9 0.8 0.7 0.6 0.5 0.4

Query Set Classification Cover Page Assumption and Top N Results Satisfied (nrs) Evidence 60% 50% 40% 30% 20% 10% While performing Navigationala navigational Informational type & Transactional search request, user tend to click only the first few URLs in the result list. 0% 0.9 0.8 0.7 0.6 0.5

Query Set Classification Click Distribution Evidence Proposed by Lee (Lee, 2005). Also based on clickthrough information. Users tend to click the same result while proposing a same navigational type query Less than 5% informational / Transactional queries CD value is over ½, while 51% navigational queries corresponding value is more than ½.

Query Set Classification A decision tree algorithm

Answer Annotation Navigational type query annotation Define: Click focus ClickFocus ( Query q, Result r) = # ( Session of q that # ( Session of clicks q) r) Annotate q with the result r whose ClickFocus value is the largest.

Answer Annotation Annotation Algorithm For a given Query Q in the Query Set and its clicked result list r1, r2,, rm : IF Q is navigational Find R in r1, r2,, rm, ClickFocus(Q,R) = ClickDistribution(Q); IF CD(Q) > T1 Annotate Q with R; EXIT; ELSE Q cannot be annotated; END IF ELSE //Q is informational Q cannot be annotated; END IF

Experiment Results Experiment data Collected by Sogou.com from Jun 2006 to Jan 2007. Over 700 million querying or clicking events totally. Annotation experiment results 5% of all results are checked mannually. #(Annotated queries) #(Checked sample set) Accuracy Jun. 06 - Aug. 06 13,902 695 98.13% Sept.06 - Nov. 06 13,884 694 97.41% Dec. 06 - Jan. 07 11,296 565 96.64%

Experiment Results Performance evaluation experiment 320 manual-developed queries and corresponding answers are used in the evaluation experiment. Correlation value between MRRs of the manual and the automatically methods is 0.965.

Applications and Future works Choosing the correct search portal Overall performance Performance for queries in a certain field Search engine monitoring Complicated computer cluster systems are used in modern search engines To notify the engineers when the search engine fails. (performance going down)

Thank you! Questions or comments?