Learning Ranking Functions with Implicit Feedback

Similar documents
Learning Ranking Functions with SVMs

Learning Ranking Functions with SVMs

Adaptive Search Engines Learning Ranking Functions with SVMs

Evaluating the Accuracy of. Implicit feedback. from Clicks and Query Reformulations in Web Search. Learning with Humans in the Loop

Machine Learning for Information Discovery

Optimizing Search Engines using Click-through Data

Large-Scale Validation and Analysis of Interleaved Search Evaluation

Learning with Humans in the Loop

Learning to Rank (part 2)

Advanced Topics in Information Retrieval. Learning to Rank. ATIR July 14, 2016

IMAGE RETRIEVAL SYSTEM: BASED ON USER REQUIREMENT AND INFERRING ANALYSIS TROUGH FEEDBACK

Learning Socially Optimal Information Systems from Egoistic Users

Ranking and Learning. Table of Content. Weighted scoring for ranking Learning to rank: A simple example Learning to ranking as classification.

Learning to Rank. Tie-Yan Liu. Microsoft Research Asia CCIR 2011, Jinan,

Evaluating search engines CE-324: Modern Information Retrieval Sharif University of Technology

Classification. 1 o Semestre 2007/2008

Optimizing Search Engines using Clickthrough Data

WebSci and Learning to Rank for IR

Information Retrieval. CS630 Representing and Accessing Digital Information. What is a Retrieval Model? Basic IR Processes

Evaluating Retrieval Performance using Clickthrough Data

Classification: Feature Vectors

Text classification II CE-324: Modern Information Retrieval Sharif University of Technology

Machine Learning. Support Vector Machines. Fabio Vandin November 20, 2017

Multimedia Information Systems

Announcements. CS 188: Artificial Intelligence Spring Classification: Feature Vectors. Classification: Weights. Learning: Binary Perceptron

Learning to Rank for Faceted Search Bridging the gap between theory and practice

Minimally Invasive Randomization for Collecting Unbiased Preferences from Clickthrough Logs

Project Presentation. Pattern Recognition. Under the guidance of Prof. Sumeet Agar wal

Development in Object Detection. Junyuan Lin May 4th

CSE 573: Artificial Intelligence Autumn 2010

Ryen W. White, Matthew Richardson, Mikhail Bilenko Microsoft Research Allison Heath Rice University

Support vector machines

CSCI 599: Applications of Natural Language Processing Information Retrieval Evaluation"

Estimating Credibility of User Clicks with Mouse Movement and Eye-tracking Information

Feature Extractors. CS 188: Artificial Intelligence Fall Some (Vague) Biology. The Binary Perceptron. Binary Decision Rule.

Automatic Identification of User Goals in Web Search [WWW 05]

Learning Dense Models of Query Similarity from User Click Logs

CS 229 Midterm Review

Chapter 8. Evaluating Search Engine

Announcements. CS 188: Artificial Intelligence Spring Generative vs. Discriminative. Classification: Feature Vectors. Project 4: due Friday.

Ranked Retrieval. Evaluation in IR. One option is to average the precision scores at discrete. points on the ROC curve But which points?

Learning to Rank. Kevin Duh June 2018

An Overview of Search Engine. Hai-Yang Xu Dev Lead of Search Technology Center Microsoft Research Asia

Fall Lecture 16: Learning-to-rank

Retrieval Evaluation

1 Case study of SVM (Rob)

Introduction to Support Vector Machines

CPSC 340: Machine Learning and Data Mining. Ranking Fall 2016

CSCI 5417 Information Retrieval Systems. Jim Martin!

Topics in Machine Learning

Support Vector Machines and their Applications

Mining User Preference Using Spy Voting for Search Engine Personalization

CS 6604: Data Mining Large Networks and Time-Series

Data Mining: Concepts and Techniques. Chapter 9 Classification: Support Vector Machines. Support Vector Machines (SVMs)

Evaluation. David Kauchak cs160 Fall 2009 adapted from:

Learning to Align Sequences: A Maximum-Margin Approach

Information Retrieval

A General Greedy Approximation Algorithm with Applications

Collaborative Metric Learning

Lecture 9: Support Vector Machines

Kernels + K-Means Introduction to Machine Learning. Matt Gormley Lecture 29 April 25, 2018

Information Retrieval

CS535 Big Data Fall 2017 Colorado State University 10/10/2017 Sangmi Lee Pallickara Week 8- A.

Lecture #11: The Perceptron

Application of Support Vector Machine Algorithm in Spam Filtering

Visual Recognition: Examples of Graphical Models

CS570: Introduction to Data Mining

Search Engines and Learning to Rank

Combine the PA Algorithm with a Proximal Classifier

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

Python & Web Mining. Lecture Old Dominion University. Department of Computer Science CS 495 Fall 2012

CS47300: Web Information Search and Management

Apache Solr Learning to Rank FTW!

Support vector machine (II): non-linear SVM. LING 572 Fei Xia

10601 Machine Learning. Model and feature selection

Link Mining & Entity Resolution. Lise Getoor University of Maryland, College Park

CPSC 340: Machine Learning and Data Mining. More Linear Classifiers Fall 2017

Introduction to Information Retrieval

Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University Infinite data. Filtering data streams

Math 454 Final Exam, Fall 2005

CS446: Machine Learning Fall Problem Set 4. Handed Out: October 17, 2013 Due: October 31 th, w T x i w

Where we are. Exploratory Graph Analysis (40 min) Focused Graph Mining (40 min) Refinement of Query Results (40 min)

DM6 Support Vector Machines

CS6375: Machine Learning Gautam Kunapuli. Mid-Term Review

Machine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013

Search Engines Chapter 8 Evaluating Search Engines Felix Naumann

Syllabus. 1. Visual classification Intro 2. SVM 3. Datasets and evaluation 4. Shallow / Deep architectures

Text Classification and Clustering Using Kernels for Structured Data

A Dendrogram. Bioinformatics (Lec 17)

Transductive Learning: Motivation, Model, Algorithms

Detecting Malicious URLs. Justin Ma, Lawrence Saul, Stefan Savage, Geoff Voelker. Presented by Gaspar Modelo-Howard September 29, 2010.

Web Search and Text Mining. Lecture 8

Spying Out Accurate User Preferences for Search Engine Adaptation

CSCI 599: Applications of Natural Language Processing Information Retrieval Retrieval Models (Part 1)"

Decision Trees: Representation

CSEP 573: Artificial Intelligence

Fractional Similarity : Cross-lingual Feature Selection for Search

SIGIR 2003 Workshop Report: Implicit Measures of User Interests and Preferences

Search Engine Architecture. Hongning Wang

Experiment Design and Evaluation for Information Retrieval Rishiraj Saha Roy Computer Scientist, Adobe Research Labs India

Transcription:

Learning Ranking Functions with Implicit Feedback CS4780 Machine Learning Fall 2011 Pannaga Shivaswamy Cornell University These slides are built on an earlier set of slides by Prof. Joachims.

Current Search Engines One-size-fits-all Hand-tuned retrieval function Hypothesis Adaptive Search Engines Different users need different retrieval functions Different collections need different retrieval functions Machine Learning Learn improved retrieval functions User Feedback as training data

Overview How can we get training data for learning improved retrieval functions? Explicit vs. implicit feedback Absolute vs. relative feedback Learning to rank algorithms: Ranking Support Vector Machine Online learning to rank algorithm with interactive feedback

Explicit Feedback Overhead for user Only few users give feedback => not representative Implicit Feedback Sources of Feedback Queries, clicks, time, mousing, scrolling, etc. No Overhead More difficult to interpret

Types of Feedback Absolute Feedback Feedback about relevance of document on absolute scale Examples Document d i is relevant to query q Document d j is not relevant to query q Document d l has relevance 0.73 with respect to query q Relative Feedback Feedback reflects preference between documents Examples Document d i is more relevant to query q than document d j Document d i is the least relevant to query q among {d i, d j, d l, d m }

Feedback from Clickthrough Data Relative Feedback: Clicks reflect preference between observed links. Absolute Feedback: The clicked links are relevant to the query. (3 < 2), (7 < 2), (7 < 4), (7 < 5), (7 < 6) 1. Kernel Machines http://svm.first.gmd.de/ 2. Support Vector Machine http://jbolivar.freeservers.com/ 3. SVM-Light Support Vector Machine http://ais.gmd.de/~thorsten/svm light/ 4. An Introduction to Support Vector Machines http://www.support-vector.net/ 5. Support Vector Machine and Kernel... References http://svm.research.bell-labs.com/svmrefs.html 6. Archives of SUPPORT-VECTOR-MACHINES... http://www.jiscmail.ac.uk/lists/support... 7. Lucent Technologies: SVM demo applet http://svm.research.bell-labs.com/svt/svmsvt.html 8. Royal Holloway Support Vector Machine http://svm.dcs.rhbnc.ac.uk Rel(1), NotRel(2), Rel(3), NotRel(4), NotRel(5), NotRel(6), Rel(7)

Strategies for Generating Relative Feedback Strategies Click > Skip Above (3>2), (5>2), (5>4) Last Click > Skip Above (5>2), (5>4) Click > Earlier Click (3>1), (5>1), (5>3) Click > Skip Previous (3>2), (5>4) Click > Skip Next (1>2), (3>4), (5>6) 1. Kernel Machines http://www.kernel-machines.org/ 2. Support Vector Machine http://jbolivar.freeservers.com/ 3. SVM-Light Support Vector Machine http://ais.gmd.de/~thorsten/svm light/ 4. An Introduction to SVMs http://www.support-vector.net/ 5. Support Vector Machine and... http://svm.bell-labs.com/svmrefs.html 6. Archives of SUPPORT-VECTOR... http://www.jisc.ac.uk/lists/support... 7. Lucent Technologies: SVM demo applet http://svm.bell-labs.com/svmsvt.html 8. Royal Holloway SVM http://svm.dcs.rhbnc.ac.uk 9. SVM World http://www.svmworld.com 10. Fraunhofer FIRST SVM page http://svm.first.gmd.de

Learning Retrieval Functions with SVMs Form of retrieval function: query q, documents d 1,,d k sort by score(q,d i ) = w 1 * (#of query words in title of d i ) + w 2 * (#of query words in anchor) + + w n * (page-rank of d i ) = w * Φ(q,d i ) Absolute Feedback: Binary Classification SVM Rel vs. NotRel Learn score so that score(q,d i )>0, if Rel score(q,d i )<0, if NotRel Relative Feedback: Ranking SVM Learn score so that score(q,d i ) > score(q,d j ) ó d i more relevant than d j

Learning Retrieval Functions from Pairwise Preferences Idea: Learn a ranking function, so that number of violated pair-wise training preferences is minimized. Form of Ranking Function: sort by rsv(q,d i ) = w 1 * (#of query words in title of d i ) + w 2 * (#of query words in anchor) + + w n * (page-rank of d i ) = w * Φ(q,d i ) Training: Select w so that if user prefers d i to d j for query q, then rsv(q, d i ) > rsv(q, d j )

Ranking Support Vector Machine Find ranking function with low error and large margin Properties Convex quadratic program Non-linear functions using Kernels Implemented as part of SVM-light http://svmlight.joachims.org 1 4 2 3

Experiment Meta-Search Engine Striver Implemented meta-search engine on top of Google, MSNSearch, Altavista, Hotbot, Excite Retrieve top 100 results from each search engine Re-rank results with learned ranking functions Experiment Setup User study on group of ~20 German machine learning researchers and students => homogeneous group of users Asked users to use the system like any other search engine Train ranking SVM on 3 weeks of clickthrough data Test on 2 following weeks

Which Ranking Function is Better? 1. Kernel Machines http://svm.first.gmd.de/ 2. Support Vector Machine http://jbolivar.freeservers.com/ 3. An Introduction to Support Vector Machines http://www.support-vector.net/ 4. Archives of SUPPORT-VECTOR-MACHINES... http://www.jiscmail.ac.uk/lists/support... 5. SVM-Light Support Vector Machine http://ais.gmd.de/~thorsten/svm light/ Google 1. Kernel Machines http://svm.first.gmd.de/ 2. Support Vector Machine http://jbolivar.freeservers.com/ 3. SVM-Light Support Vector Machine http://ais.gmd.de/~thorsten/svm light/ 4. An Introduction to Support Vector Machines http://www.support-vector.net/ 5. Support Vector Machine and Kernel... References http://svm.research.bell-labs.com/svmrefs.html 6. Archives of SUPPORT-VECTOR-MACHINES... http://www.jiscmail.ac.uk/lists/support... 7. Lucent Technologies: SVM demo applet http://svm.research.bell-labs.com/svt/svmsvt.html 8. Royal Holloway Support Vector Machine http://svm.dcs.rhbnc.ac.uk 1. Kernel Machines http://svm.first.gmd.de/ 2. SVM-Light Support Vector Machine http://ais.gmd.de/~thorsten/svm light/ 3. Support Vector Machine and Kernel... References http://svm.research.bell-labs.com/svmrefs.html 4. Lucent Technologies: SVM demo applet http://svm.research.bell-labs.com/svt/svmsvt.html 5. Royal Holloway Support Vector Machine http://svm.dcs.rhbnc.ac.uk Learned Approach Experiment setup generating unbiased clicks for fair evaluation. Validity Clickthrough in combined ranking gives same results as explicit feedback under mild assumptions [Joachims, 2003].

Results Ranking A Ranking B A better B better Tie Total Learned Google 29 13 27 69 Learned MSNSearch 18 4 7 29 Learned Toprank 21 9 11 41 Result: Learned > Google Learned > MSNSearch Learned > Toprank

Learned Weights Weight Feature 0.60 cosine between query and abstract 0.48 ranked in top 10 from Google 0.24 cosine between query and the words in the URL 0.24 doc ranked at rank 1 by exactly one of the 5 engines... 0.22 host has the name citeseer 0.17 country code of URL is ".de" 0.16 ranked top 1 by HotBot... -0.15 country code of URL is ".fi" -0.17 length of URL in characters -0.32 not ranked in top 10 by any of the 5 search engines -0.38 not ranked top 1 by any of the 5 search engines

Online learning with interactive feedback Idea: Learn a ranking function online based on interaction from the user Present a ranking à Observe user feedback à Update y ȳ Goal: Present good rankings to a user Advantages of online learning to rank: ü Naturally suited to how a user interacts ü No need of an expensive offline training data ü Theoretical guarantees ü Performs well compared to offline learning

Listwise feedback (1,3,7,2,4,5,6,8) > (1,2,3,4,5,6,7,8) 1. Kernel Machines http://svm.first.gmd.de/ 2. Support Vector Machine http://jbolivar.freeservers.com/ 3. SVM-Light Support Vector Machine http://ais.gmd.de/~thorsten/svm light/ 4. An Introduction to Support Vector Machines http://www.support-vector.net/ 5. Support Vector Machine and Kernel... References http://svm.research.bell-labs.com/svmrefs.html 6. Archives of SUPPORT-VECTOR-MACHINES... http://www.jiscmail.ac.uk/lists/support... 7. Lucent Technologies: SVM demo applet http://svm.research.bell-labs.com/svt/svmsvt.html 8. Royal Holloway Support Vector Machine http://svm.dcs.rhbnc.ac.uk

Learning Model Utility of a ranking to a query as a linear model: w Φ(q t, y t ) Optimal ranking for a query: α informative feedback: y t arg max y Y w Φ(q t, y) w Φ(q t, ȳ t ) w Φ(q t, y t ) α(w Φ(q t, y t ) w Φ(q t, y t ))

User Feedback w Φ(q t, ȳ t ) w Φ(q t, y t ) α(w Φ(q t, y t ) w Φ(q t, y t )) y αl ȳ y l

Interactive Perceptron Algorithm Initialize w 1 0 for t =1to T do Observe q t Present y t arg max y Y w t Φ(q t, y) Receive feedback ȳ t Update w t+1 w t + Φ(q t, ȳ t ) Φ(q t, y t ) end for

Notion of regret: 1 T T t=1 Theoretical Guarantee w (Φ(q t, y t ) Φ(q t, y t )) Conditions: α informative feedback boundedness of features Φ(q, y) 2 R 2 Guarantee: Regret 2Rw α T

w T +1w T +1 Proof =w T w T +2w T (Φ(q T, ȳ T ) Φ(q T, y T )) 0 +(Φ(q T, ȳ T ) Φ(q T, y T )) (Φ(q T, ȳ T ) Φ(q T, y T )) wt w T +4R 2 wt 1w T 1 +2.4R 2 w 1 w 1 +4R 2 T 4R 2 T w T +1 2R T Recall : w t+1 w t + Φ(q t, ȳ t ) Φ(q t, y t ) Φ(q, y) 2 R 2 w 1 =0

From the assumption on feedback w T +1w = Therefore, Recall : T w (Φ(q t, ȳ t ) Φ(q t, y t )) T α w (Φ(q t, yt ) Φ(q t, y t )) t=1 T α w (Φ(q t, yt ) Φ(q t, y t )) w w T +1 t=1 t=1 ww T +1 2R T w w w T +1 ww T +1 w T +1 2R T

Rearranging 1 T T t=1 w (Φ(q t, y t ) Φ(q t, y t )) 2Rw α T As T average regret approaches 0

Experiment Dataset: Yahoo! Learning to rank dataset Utility of a ranking: w Φ(q, y) = w Baseline: 5 i=1 w φ(q,d yi ) log(1 + i) best least squares fit on the data Pairwise SVM trained repeatedly (query,url) relevance (q_1,d_1) 0 (q_1,d_2) 3 (q_1,d_3) 4 (q_2,d_1) 1 (q_2,d_2) 1 (q_2,d_3) 0 (q_2,d_4) 2

Clean Feedback 1.4 1.2 1 avg. util regret 0.8 0.6 0.4 0.2 0 10 0 10 1 10 2 10 3 10 4 t

Noisy Feedback 1.8 SVM Online Learning 1.6 1.4 avg. util regret 1.2 1 0.8 0.6 0.4 10 0 10 1 10 2 10 3 10 4 t

Conclusions Clickthrough data can provide accurate feedback Clickthrough provides relative instead of absolute judgments Ranking SVM can learn effectively from relative preferences Improved retrieval through personalization in meta search Online learning is very natural in this setting Peforms better than repeatedly training an SVM Current and future work Applying the online learning model to a realworld setup Diversity in search results via online learning Better theoretical guarantees for particular families of loss functions Incorporating similarity information across users