Towards Predicting Web Searcher Gaze Position from Mouse Movements

Similar documents
In the Mood to Click? Towards Inferring Receptiveness to Search Advertising

Visualization of User Eye Movements for Search Result Pages

Estimating Credibility of User Clicks with Mouse Movement and Eye-tracking Information

Learning Ranking Functions with SVMs

Machine Learning Applications to Modeling Web Searcher Behavior Eugene Agichtein

No Clicks, No Problem: Using Cursor Movements to Understand and Improve Search

How Do Sponsored Search Results Affect User Behavior in Web Search?

Evaluating the Accuracy of. Implicit feedback. from Clicks and Query Reformulations in Web Search. Learning with Humans in the Loop

ViewSer: Enabling Large-Scale Remote User Studies of Web Search Examination and Interaction

Adaptive Search Engines Learning Ranking Functions with SVMs

Measurement and Modeling of Eye-mouse Behavior in the Presence of Nonlinear Page Layouts

Exploring Client-Side Instrumentation for Personalized Search Intent Inference: Preliminary Experiments

Learning Ranking Functions with SVMs

Modeling Contextual Factors of Click Rates

Different Users, Different Opinions: Predicting Search Satisfaction with Mouse Movement Information

A Task Level Metric for Measuring Web Search Satisfaction and its Application on Improving Relevance Estimation

Dissertation Research Proposal

Modeling User Behavior and Interactions. Lecture 2: Interpreting Behavior Data. Eugene Agichtein

Evaluating Search Engines by Modeling the Relationship Between Relevance and Clicks

Analysis of User Behavior for Web Search Success using Eye Tracker data

An Analysis Study on How Female College Students View the Web Search Results Using Eye Tracking Methodology

Detecting Good Abandonment in Mobile Search

Data Mining. 3.5 Lazy Learners (Instance-Based Learners) Fall Instructor: Dr. Masoud Yaghini. Lazy Learners

Detecting Multilingual and Multi-Regional Query Intent in Web Search

Text 2.0. Keywords. General Terms Human Factors, Languages, Experimentation

Generating and Using Gaze-Based Document Annotations

Improving Stack Overflow Tag Prediction Using Eye Tracking Alina Lazar Youngstown State University Bonita Sharif, Youngstown State University

Advanced Search Techniques for Large Scale Data Analytics Pavel Zezula and Jan Sedmidubsky Masaryk University

Preliminary Evidence for Top-down and Bottom-up Processes in Web Search Navigation

A Model to Estimate Intrinsic Document Relevance from the Clickthrough Logs of a Web Search Engine

Interaction Model to Predict Subjective Specificity of Search Results

Influence of Vertical Result in Web Search Examination

IJREAT International Journal of Research in Engineering & Advanced Technology, Volume 1, Issue 5, Oct-Nov, ISSN:

Automatic Identification of User Goals in Web Search [WWW 05]

Effect of log-based Query Term Expansion on Retrieval Effectiveness in Patent Searching

Classifying Depositional Environments in Satellite Images

STATISTICS (STAT) Statistics (STAT) 1

Letter Pair Similarity Classification and URL Ranking Based on Feedback Approach

Domain Specific Search Engine for Students

Sentiment analysis under temporal shift

Inferring Searcher Intent

A Dynamic Bayesian Network Click Model for Web Search Ranking

Fighting Search Engine Amnesia: Reranking Repeated Results

Problem 1: Complexity of Update Rules for Logistic Regression

Evaluation Metrics. (Classifiers) CS229 Section Anand Avati

ResPubliQA 2010

CS 572: Information Retrieval. Implicit Searcher Feedback

Predicting Messaging Response Time in a Long Distance Relationship

NUS-I2R: Learning a Combined System for Entity Linking

Interpreting User Inactivity on Search Results

Large-Scale Analysis of Individual and Task Differences in Search Result Page Examination Strategies

A New Technique to Optimize User s Browsing Session using Data Mining

This research aims to present a new way of visualizing multi-dimensional data using generalized scatterplots by sensitivity coefficients to highlight

Improving Search Result Summaries by Using Searcher Behavior Data

Predicting Popular Xbox games based on Search Queries of Users

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016

CS229 Final Project: Predicting Expected Response Times

Mining the Search Trails of Surfing Crowds: Identifying Relevant Websites from User Activity Data

5. search engine marketing

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

Time-Aware Click Model

Applying the KISS Principle for the CLEF- IP 2010 Prior Art Candidate Patent Search Task

Topics in Machine Learning

CURIOUS BROWSERS: Automated Gathering of Implicit Interest Indicators by an Instrumented Browser

Diversification of Query Interpretations and Search Results

Personalized Information Retrieval. Elena Holobiuc Iulia Pasov Alexandru Agape Octavian Sima Bogdan Cap-Bun

A Comparison of Error Metrics for Learning Model Parameters in Bayesian Knowledge Tracing

Modern Retrieval Evaluations. Hongning Wang

Accelerometer Gesture Recognition

Success Index: Measuring the efficiency of search engines using implicit user feedback

CS6716 Pattern Recognition

Supporting Orientation during Search Result Examination

Success Index: Measuring the efficiency of search engines using implicit user feedback

Adaptive Robotics - Final Report Extending Q-Learning to Infinite Spaces

Deep Face Recognition. Nathan Sun

CSE 573: Artificial Intelligence Autumn 2010

A user-oriented webpage ranking algorithm based on user attention time

A Task-Based Evaluation of an Aggregated Search Interface

S-MART: Novel Tree-based Structured Learning Algorithms Applied to Tweet Entity Linking

CSI5387: Data Mining Project

Task Description: Finding Similar Documents. Document Retrieval. Case Study 2: Document Retrieval

Pixels to Voxels: Modeling Visual Representation in the Human Brain

MetaData for Database Mining

On the automatic classification of app reviews

Extracting Visual Snippets for Query Suggestion in Collaborative Web Search

CS/NEUR125 Brains, Minds, and Machines. Due: Wednesday, April 5

Attention modulates spatial priority maps in human occipital, parietal, and frontal cortex

Query Refinement and Search Result Presentation

Rank Measures for Ordering

CS 6740: Advanced Language Technologies April 2, Lecturer: Lillian Lee Scribes: Navin Sivakumar, Lakshmi Ganesh, Taiyang Chen.

PERSONALIZED MOBILE SEARCH ENGINE BASED ON MULTIPLE PREFERENCE, USER PROFILE AND ANDROID PLATFORM

Predicting Query Performance on the Web

Eyetracking Study: Kent State University Library. Janie Ralston

Ghent University-IBCN Participation in TAC-KBP 2015 Cold Start Slot Filling task

Efficient Multiple-Click Models in Web Search

Comment Extraction from Blog Posts and Its Applications to Opinion Mining

Mini-project 2 CMPSCI 689 Spring 2015 Due: Tuesday, April 07, in class

Fast or furious? - User analysis of SF Express Inc

Nearest Neighbor Predictors

The Effects of Semantic Grouping on Visual Search

Transcription:

Towards Predicting Web Searcher Gaze Position from Mouse Movements Qi Guo Emory University 400 Dowman Dr., W401 Atlanta, GA 30322 USA qguo3@emory.edu Eugene Agichtein Emory University 400 Dowman Dr., W401 Atlanta, GA 30322 USA eugene@mathcs.emory.edu Copyright is held by the author/owner(s). CHI 2010, April 10 15, 2010, Atlanta, Georgia, USA. ACM 978-1-60558-930-5/10/04. Abstract A key problem in information retrieval is inferring the searcher s interest in the results, which can be used for implicit feedback, query suggestion, and result ranking and summarization. One important indicator of searcher interest is gaze position that is, the results or the terms in a result listing where a searcher concentrates her attention. Capturing this information normally requires eye tracking equipment, which until now has limited the use of gaze-based feedback to the laboratory. While previous research has reported a correlation between mouse movement and gaze position, we are not aware of prior work on automatically inferring searcher s gaze position from mouse movement or similar interface interactions. In this paper, we report the first results on automatically inferring whether the searcher s gaze position is coordinated with the mouse position a crucial step towards predicting the searcher gaze position by analyzing the computer mouse movements. Keywords Web search, searcher behavior, eye-mouse coordination ACM Classification Keywords H.3.3 [Information Storage and Retrieval]: Information Search and Retrieval search and selection process. General Terms Design, Experimentation, Human Factors 3601

Introduction A key problem in web search is inferring searcher intent and interest in the results, which is important for tasks such as implicit feedback [1], query suggestion, and result ranking and summarization. An accurate indicator of searcher interest is gaze position [1], which normally requires eye tracking equipment. Unfortunately, eye tracking-based feedback is not yet feasible on web-scale. However, work by Rodden et al. [10] has shown that sometimes there is a correlation between mouse and eye position, and described general patterns of mouse/eye coordination. Yet, we are not aware of any reports of automatically inferring the gaze position from mouse movements and other interaction features. Furthermore, it has not been shown if it is possible to predict, for a given point in the interaction process, whether the gaze position and mouse position are coordinated -- which is the first step towards predicting the searcher s gaze position. Thus, we explore the basic research question: Is it possible to predict the precise times when mouse and gaze position are closely coordinated, based only on mouse position and movement? Our work advances the state of the art by reporting, to our knowledge, the first successful method for inferring the times of mouse-eye coordination from interface interactions alone. Specifically, we examine the characteristics of eye-mouse coordination (measured as Euclidian distance between the mouse and gaze position), using data collected from a controlled user study of web search with 10 subjects and 200 instances of search tasks. We consider factors such as basic search task type and result quality that may have impact on the eye-mouse coordination patterns. Finally, we use machine learning techniques to infer mouse-eye coordination, using only the features of the mouse movement. Our results demonstrate that it is indeed possible to predict with over 70% accuracy whether gaze position and mouse position are coordinated. While this is ongoing work, our results may be already useful for search tasks such as implicit feedback, query suggestion, and click prediction. Related work Recently, eye tracking has started to emerge as a useful technology for understanding searcher behavior (e.g., [3, 7]). However, eye tracking equipment is not usually available for web search users, and there has been significant interest in inferring searcher interest from clicks and other interface actions [8, 9]. In particular, searcher interactions have been used to help disambiguate query intent [5], for implicit feedback [2], and for more accurate prediction of search result clickthrough [6]. However, the work above side-steps the problem of inferring user interest as expressed by gaze position, and focuses on application-specific measures. In contrast, we are interested in the fundamental problem of predicting mouse-eye coordination, which could have applications beyond our specific web search setting. We significantly expand on the previous work by Rodden et al. [10] which described general patterns of eye-mouse coordination, and reported tantalizing statistics of overall correlation between mouse and eye position. In contrast, we present, to our knowledge, the first method to predict the gaze position from the mouse movements. Study Apparatus To capture the user interactions, including mouse movements, we use our extension of a LibX plugin for the Firefox browser. The events are encoded in a string and sent to the server as HTTP requests for analysis. We sample mouse movements at every 5 pixels moved, or every 50 ms, whichever is more frequent. Eye tracking was performed using the Tobii T120 system, paired with a 17 LCD monitor set at the resolution of 1280 1024 for all experiments. The Web browser was maximized for all subjects, allowing us to reproduce exactly the same result presentation for all subjects. 3602

Figure 1: Eye-mouse distance (X, Y, and Euclidean). Participants Ten subjects were recruited for this user study. All the subjects were graduate students and staff in the Emory Math & CS department; as such, all subjects were technically savvy and had some experience with web search. Tasks and Search Results We used a set of 20 web search tasks, sampled from the web search usage logs collected from public-use machines in the Emory University Libraries. Table 1 reports example search tasks, together with the descriptions provided to the participants. Half of the tasks were Navigational (find a particular web site or page) and the others were Informational (a specific piece of information on any web page). Query Navigational Tasks cnn wikipedia Informational Tasks circuit city ceo 1995 setting up a gym in your house Description Find the official site of CNN Find the official site of Wikipedia Find out who was the CEO of circuit city in 1995 Find out how to set up a home gym or work out area Table 1: Example search tasks (queries and descriptions). For each task, a pre-specified query was provided and the associated first Search Engine Result Page (SERP) was stored by querying the Google search engine. To ensure result consistency, only the stored copies were shown to the participants. Nine of the 20 SERPs contained images while the rest do not. To explore the effects of result quality, the search results for half of the tasks (namely, for 5 navigational and 5 informational tasks) were randomized. Thus, our dataset consisted of 20 SERPs, with 10 original result ranking, and 10 with the top search results in random order. Procedure To begin a task, the participants were presented a list of queries and task descriptions. For each task, the participants were instructed to find the most relevant result on the cached first Search Engine Result Page (SERP) for the pre-specified query. Once they reached the SERP, they could read the text and find the most relevant result to click. Clicking a chosen result recorded the subjects decision, and redirected them back to the task description list to start the next task. Characterizing Eye-Mouse Coordination Eye-mouse distance characteristics In order to calculate the set of distances between mouse and eye, we matched each mouse point with the eye fixation (if any) in the same time interval. The overall distributions of distances are reported in Figure 1, and are similar to Rodden et al. [10]. The mean Euclidean distance is 225 pixels (σ=178); the median is 180. For the Y direction there is a higher peak around the origin than for the X direction, suggesting that mouse and eye positions coordinate more closely in the vertical direction, which agrees with the findings in [10]. Factors X direction Y direction Euclidean All 166 115 228 Task Navigational 147 ** 124 ** 219 Ranking Informational 183 ** 107 ** 235 Original 152 * 119 219 Randomized 179 * 111 236 Image? Yes 159 116 220 No 173 114 234 Table 2. Mean eye-mouse distance for varying search factors. Eye-mouse coordination for varying factors We now analyze the eye-mouse distance according to the task (Informational vs. Navigational), Ranking (Original vs. Randomized) and Result type (Image or No Image). The 3603

corresponding means for each pair of variables are summarized in Table 2. All significant differences with twotailed unpaired t-tests for each variable (i.e., task, result order, image in result) are marked using * for p <.1 and ** for p <.05. As Table 2 reports, the mean Euclidean distance is relatively consistent across different factors, while X and Y distance exhibit significant differences for task type (p<0.05). For navigational tasks, the mean X distance is smaller, while the mean Y distance is higher - which could be explained by easier to find results by scanning the result page without moving the mouse, at least in the original ranking. Interestingly, low result quality (simulated by randomized ranking) forced the searchers to examine the results more carefully. We posit that it is in these cases, predicting gaze position and identifying the few relevant results could be particularly helpful. Finally, the presence of images did not have a significant effect on overall mouse-eye coordination. Eye-mouse coordination patterns Previous work has identified three patterns of mouse-eye coordination in search, namely incidental mouse usage, bookmarking, and follow [10]. The follow and bookmarking patterns are particularly important as they could potentially be used to provide evidence for implicit feedback. We illustrate these patterns in Figure 3 (A-F). In panels A, C, and E, the eye fixations (blue circles) and mouse positions (black circles) are overlaid on the original SERPs. Panels B, D, and F plot the corresponding eye-mouse Euclidian distance values over time. A B Eye-mouse coordination for different subjects The mean eye-mouse distances for each subject across all tasks are reported in Figure 2. Four of the participants exhibit significant differences in mean mouse-eye distance (p<0.01, marked by * on the X axis). However, most subjects maintain roughly 200 pixel eye-mouse distance on average. Further investigation with a larger participant group, is needed to characterize individual differences between subjects. C D E F Figure 2. Eye-mouse Euclidean distance for different participants (*=p<.01). The dashed line represents the mean distance over all participants. Figure 3. Examples of mouse-eye coordination and corresponding Euclidian distance time series: incidental mouse usage (A-B), bookmarking (C-D), and follow (E-F). 3604

Feature Description x i y i t i X coordinate Y coordinate time elapsed t i t i t i-1 x i x i x i-1 y i y i y i-1 vh i abs( x i/ t i) dh i sign( x i/ t i) vv i abs( y i/ t i) dv i sign( y i/ t i) pid Participant ID Table 3. Features for representing a mouse sample point i The distance patterns exhibited by the incidental (A-B) and bookmarking (C-D) mouse usage illustrate the difficulty in automatically detecting eye-mouse coordination: in both cases, the mouse is stationary for a long period, indicated by a relatively larger black circle. However, in the incidental case the mouse position is meaningless, while in the bookmarking case, the mouse position indicates a result of interest which has been carefully examined. One possible distinguishing characteristic is the X-position of the mouse, which is one of the features we use for prediction. Predicting Eye-Mouse Coordination We formulate our prediction task as binary classification: that is, we classify each mouse sample point into two classes: InFocus (eye-mouse distance is below a threshold of T pixels) and Away (eye-mouse distance is above a threshold of T pixels). We experimented with T values from 200 to 100 pixels, that is, from roughly the mean Euclidean eye-mouse distance to roughly the vertical size of one search result. Evaluation Metrics We use standard IR and classification metrics: Accuracy (Acc): Fraction of points in the timeline that were correctly predicted, reported in %. Precision (Prec): For a given class, fraction of predicted instances that were correctly predicted. Recall (Rec): For a given class, fraction of all true instances that were correctly predicted. Macro-averaged F1 (F1): F1 measure, computed as 2*Prec*Rec / (Prec + Rec), averaged across both classes. Note that we are particularly interested in the InFocus predictions, as the most useful for implicit feedback and other applications. Thus, we explicitly report the Precision and Recall for the InFocus class in addition to the overall Accuracy and F1 performance. Baseline method To gauge the performance of our classification methods, we also report performance of a naive baseline, which always guesses the majority class (e.g., the Away class for threshold = 100, which is true for almost 74% of the data). Thus, our majority baseline has significantly higher accuracy than a more traditional random baseline (50% accuracy). Our Model and Implementation We represent each mouse sample point as a feature vector, with values corresponding to the X-coordinate, Y-coordinate, time elapsed since page load, velocity, and others (see Table 3, sidebar, for the full list). We also include a feature for the participant ID to allow for the possibility of personalizing the predictions. We experimented with different classification algorithms, including Support Vector Machines (SVMs), Random Forrest (RF), decision trees, and boosting-based algorithms. For this setup, logistic regression-based classifiers, specifically the LogitBoost (LB) algorithm [4] performed best, so we report classification results for the LB algorithm only. Experimental Setup We used four-fold cross validation: in each fold, 75% of the SERPs were used for training and the rest for testing. We report the average of the results across the folds. Five- and Three-fold cross validation setups resulted in similar classifier performance. Note that we always split the data by <SERP, Participant> pair, to avoid putting mouse points from the same SERP into both training and test sets, which could result in artificially high classification accuracy. Experimental Results We report the main experimental results in Table 4. The LB classifier (our method) significantly outperforms Baseline for the T value of 200, exhibiting the accuracy of 74% compared to the Baseline system (majority class=infocus) accuracy of 55%. Interestingly, reducing the threshold T to 150 and 100 pixels (for more fine-grained prediction), allows our system to maintain the accuracy of 73% and 77% respectively, 3605

Figure 4. Predictions (red triangles) for incidental, bookmarking, and follow mouse use examples, for T=150. The triangles above T (dotted line) are predictions for the Away class; triangles below T predict InFocus class. Blue lines show the actual eyemouse distances. Correct predictions are those on the same side of the dotted line as the actual distance. compared to the Baseline (majority class=away) with the accuracy of 58% and 74% respectively. In particular, with the distance threshold T=100 pixels, LB is able to correctly identify when the eye and mouse positions are coordinated (InFocus), with nearly 60% precision; with T=150, the Precision and Recall of predicting coordination increase to 68% and 65%, respectively. Threshold (T) Class Method Both InFocus Acc F1 Prec Rec 100 Baseline (Away) 73.8 42.5 0 0 LB 76.7 66.1 58.1 39.6 150 Baseline (Away) 58.2 36.8 0 0 LB 72.5 71.6 67.9 64.8 200 Baseline (InFocus) 55.1 35.5 55.1 100 LB 74.3 73.6 74.4 81.0 Table 4. Classification results for predicting eye-mouse distance To understand the contribution of the different features, we computed the Information Gain (IG) for each. For the prediction task with the threshold T=100, we find that x i (mouse X-coordinate) contributes the most, followed by t i (time elapsed since starting to view the results), pid (participant ID), y i (mouse Y-coordinate), t i (time since last mouse move), followed by the velocity and direction features. We illustrate prediction performance for the earlier examples in Figure 4 (sidebar). Our approach appears to be most effective for detecting the incidental and follow patterns; however, detecting the bookmarking cases may require richer feature representation, which is the subject of our future work. Conclusions and Future Work We explored the feasibility of predicting eye-mouse coordination using the features derived from mouse movements. Our experiments show that we can predict the regions where the eye and mouse position are within 100 pixels of each other, with the accuracy of nearly 77%. However, there is still much room to improve the accuracy and robustness of our prediction methods. As the natural next steps, we are exploring a richer feature representation of mouse movement and other result-page interactions. We then plan to use our predictions to identify behavioral patterns such as bookmarking. Finally, we plan to incorporate our predictions as more accurate feedback into real search tasks. Acknowledgements We thank Yahoo! Research for partially supporting this work through a faculty research grant, and Charles L. Clarke for valuable discussions. References [1] Buscher, G., Dengel A., van Elst, L. Query expansion using gaze-based feedback on the subdocument level. In SIGIR 2008 [2] Buscher, van Elst, L., and G., Dengel A. Segment-level display time as implicit feedback: a comparison to eye tracking. In SIGIR 2009 [3] Cutrell, E., and Guan, Z. What are you looking for? an eyetracking study of information usage in web search. In CHI 2007 [4] Friedman, J., Hastie, T., and Tibshirani, R. Additive logistic regression: A statistical view of boosting. In Ann. of Stat. 1998 [5] Guo, Q., and Agichtein, E. Exploring mouse movements for inferring query intent. In SIGIR 2008 [6] Guo, Q., Agichtein, E., Clarke C., and Ashkan, A. In the mood to click? Towards inferring receptiveness to search advertising. In WI 2009 [7] Joachims, T., Granka, L., Pan, B., Hembrooke, H., and Gay, G. Accurately interpreting clickthrough data as implicit feedback. In SIGIR 2005 [8] Piwowarski, B., Dupret, G., and Jones, R. Mining user web search activity with layered bayesian networks or how to capture a click in its context. In WSDM 2009 [9] Radlinski, F. and Joachims, T. Query chains: learning to rank from implicit feedback. In KDD 2005 [10] Rodden, K., Fu, X., Aula, A., and Spiro, I. Eye-mouse coordination patterns on web search results pages, In Ext. Abstracts, CHI 2008 3606