Extracting Latent User Preferences Using Sticky Sliders

Similar documents
vector space retrieval many slides courtesy James Amherst

Matrix Co-factorization for Recommendation with Rich Side Information HetRec 2011 and Implicit 1 / Feedb 23

REMOVAL OF REDUNDANT AND IRRELEVANT DATA FROM TRAINING DATASETS USING SPEEDY FEATURE SELECTION METHOD

Lecture on Modeling Tools for Clustering & Regression

CS Introduction to Data Mining Instructor: Abdullah Mueen

Query Independent Scholarly Article Ranking

Event: PASS SQL Saturday - DC 2018 Presenter: Jon Tupitza, CTO Architect

The k-means Algorithm and Genetic Algorithm

Rank Measures for Ordering

Gene Clustering & Classification

1 Case study of SVM (Rob)

Clustering. Bruno Martins. 1 st Semester 2012/2013

Chapter 2 Basic Structure of High-Dimensional Spaces

CS231A Course Project Final Report Sign Language Recognition with Unsupervised Feature Learning

Fuzzy Segmentation. Chapter Introduction. 4.2 Unsupervised Clustering.

Understanding Clustering Supervising the unsupervised

Weka ( )

Knowledge Discovery and Data Mining. Neural Nets. A simple NN as a Mathematical Formula. Notes. Lecture 13 - Neural Nets. Tom Kelsey.

Machine Learning in Biology

Lecture 10: Semantic Segmentation and Clustering

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov

Weighted Alternating Least Squares (WALS) for Movie Recommendations) Drew Hodun SCPD. Abstract

Knowledge Discovery and Data Mining

Feature selection. LING 572 Fei Xia

Feature Selection for fmri Classification

Construction Change Order analysis CPSC 533C Analysis Project

Knowledge Discovery and Data Mining 1 (VO) ( )

1 Training/Validation/Testing

Using Machine Learning to Optimize Storage Systems

CHAPTER 4 WEIGHT-BASED DESIRABILITY METHOD TO SOLVE MULTI-RESPONSE PROBLEMS

CS490W. Text Clustering. Luo Si. Department of Computer Science Purdue University

Optimizing Search Engines using Click-through Data

Text Modeling with the Trace Norm

Link Recommendation Method Based on Web Content and Usage Mining

Statistics 202: Data Mining. c Jonathan Taylor. Week 8 Based in part on slides from textbook, slides of Susan Holmes. December 2, / 1

A Collaborative User-centered Approach to Fine-tune Geospatial

over Multi Label Images

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Copyright 2007 Pearson Addison-Wesley. All rights reserved. A. Levitin Introduction to the Design & Analysis of Algorithms, 2 nd ed., Ch.

Clustering with Reinforcement Learning

such a manner that we are able to understand, grasp and grapple with the problem at hand in a more organized fashion.

CS 664 Flexible Templates. Daniel Huttenlocher

A Course in Machine Learning

Learning a Region of User s Preference for Product Recommendation Anbarasu Sekar and Sutanu Chakraborti

CSE 573: Artificial Intelligence Autumn 2010

Improving the Efficiency of Fast Using Semantic Similarity Algorithm

MASS Modified Assignment Algorithm in Facilities Layout Planning

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

A Constrained Spreading Activation Approach to Collaborative Filtering

Unsupervised Learning : Clustering

Cognitive Walkthrough. Francesca Rizzo 24 novembre 2004

Simplicial Global Optimization

Centralities (4) By: Ralucca Gera, NPS. Excellence Through Knowledge

Expectation Maximization (EM) and Gaussian Mixture Models

Scalable Clustering of Signed Networks Using Balance Normalized Cut

CSE 6242 A / CS 4803 DVA. Feb 12, Dimension Reduction. Guest Lecturer: Jaegul Choo

Collaborative Filtering for Netflix

K-means clustering Based in part on slides from textbook, slides of Susan Holmes. December 2, Statistics 202: Data Mining.

Keywords: clustering algorithms, unsupervised learning, cluster validity

Introduction to Information Retrieval

Algorithms for Grid Graphs in the MapReduce Model

BCN Decision and Risk Analysis. Syed M. Ahmed, Ph.D.

Data Partitioning. Figure 1-31: Communication Topologies. Regular Partitions

A Late Acceptance Hill-Climbing algorithm the winner of the International Optimisation Competition

Representing Symbolic Reasoning

CSE 158. Web Mining and Recommender Systems. Midterm recap

Latent Variable Models for Structured Prediction and Content-Based Retrieval

CREATIVE ASSERTION AND CONSTRAINT METHODS FOR FORMAL DESIGN VERIFICATION

CLUSTERING. CSE 634 Data Mining Prof. Anita Wasilewska TEAM 16

Ali Abur Northeastern University Department of Electrical and Computer Engineering Boston, MA 02115

Experimental Data and Training

Artificial Intelligence. Programming Styles

Gestão e Tratamento da Informação

Clustering CS 550: Machine Learning

Clustering and Dimensionality Reduction. Stony Brook University CSE545, Fall 2017

Module 1 Introduction. IIT, Bombay

Effective probabilistic stopping rules for randomized metaheuristics: GRASP implementations

Computer Vision 2 Lecture 8

Infographics and Visualisation (or: Beyond the Pie Chart) LSS: ITNPBD4, 1 November 2016

UX Research in the Product Lifecycle

SE 263 R. Venkatesh Babu. Object Tracking. R. Venkatesh Babu

HEURISTIC OPTIMIZATION USING COMPUTER SIMULATION: A STUDY OF STAFFING LEVELS IN A PHARMACEUTICAL MANUFACTURING LABORATORY

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

Variational Methods for Discrete-Data Latent Gaussian Models

7/15/2016 ARE YOUR ANALYSES TOO WHY IS YOUR ANALYSIS PARAMETRIC? PARAMETRIC? That s not Normal!

CSE 7/5337: Information Retrieval and Web Search Document clustering I (IIR 16)

Exploratory data analysis for microarrays

Random projection for non-gaussian mixture models

6. Relational Algebra (Part II)

Mining Web Data. Lijun Zhang

On Average ONLY 2% of your website visitors will actually make contact with you.

Chapter 1 Introduction to using R with Mind on Statistics

Figure : Example Precedence Graph

CHAPTER 6 PERCEPTUAL ORGANIZATION BASED ON TEMPORAL DYNAMICS

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

Applied Bayesian Nonparametrics 5. Spatial Models via Gaussian Processes, not MRFs Tutorial at CVPR 2012 Erik Sudderth Brown University

Lecture 7: Decision Trees

Analyzing Stochastic Gradient Descent for Some Non- Convex Problems

Supervised Learning. CS 586 Machine Learning. Prepared by Jugal Kalita. With help from Alpaydin s Introduction to Machine Learning, Chapter 2.

A Web Recommendation System Based on Maximum Entropy

Transcription:

Extracting Latent User Preferences Using Sticky Sliders Garrett Eastham {Stanford University} Machine learning plays a critical role in user modeling applications where dynamic systems are trained against collected user behavior data to predict future actions, such as web search. Preference elicitation remains a difficult task in the design of such systems, a task further complicated by users own ambiguous search criteria and knowledge of a problem domain. A system for preference extraction has been designed based on previous research in feature discovery interfaces and a supervised machine learning model is developed to accurately estimate a user s latent decision criteria based on limited interaction data. ABSTRACT QUERY FORMATION AND PRIOR WORK Formal study of query construction and evaluation has consistently revealed that the modern user s search criteria can be quite complex. In many scenarios, the users decision criteria are ill-formed at outset, and in fact, most have little idea about the particular distinguishing features that will guide their final selection. Cognition forces the mind to create these features during the categorization process, and during an informational discriminatory task such as search, the human mind subconsciously constructs such preferences on its own. (Schyns 1997, 1998) Further work has shown that these mental preferences are closely tied with the nature and timing of the information presented to the user, and, as a result, systems can be built that modify the end users own decision criteria for better or worse. (Pu 2003, Slovic 1995) My previous research project, called HeyElroy, involved designing and testing an interface for semantic feature discovery. The system was designed to take advantage of the previously mentioned concepts of preference construction by allowing users to explore and discover features relevant to a problem space in addition to the items of inquiry in a natural, iterative manner. (Eastham 2010) This work was built on a previous study exploring the role of ad-hoc concepts in decision making within single and group based environments, where a correlation was discovered between the confidence of selection and the number of observed items / features. CREATING TANGIBLE FEATURE-TO-FEATURE TRADEOFFS The HeyElroy system has been heavily modified to explore a new interaction paradigm I call sticky sliders that is used to generate sample data to train a preference matrix against. Figure 1 shows the primary attribute matrix that users are presented with upon loading the comparison system. Each item s attributes are listed vertically in each column and the comparable features are listed on the left in each row with interactive sliders under each feature label.

Figure 1 The sliders control linear weights on the attributes that effect their contribution to the overall rank of the column relative to the rest of the compared items. To calculate a column ranking, the columns attribute values are summed according to their weighted values and then ranked according to decreasing value: where R i is the rank of the i th column for an attribute matrix with m features and n items. Initially, the weights W are all set to the same value (arbitrarily set to integer value 50 across all features), and the position of the slider determines the value of W (all the way left corresponds to 0 and all the way right 100). Because in an ideal query, all users would want to maximize all attributes in searching for an item, the movement of sliders is constrained by a real world n x n preference matrix R, constructed: In this representation, each row R i represents the tradeoffs corresponding with changes made to the i th feature weight. The diagonal values are normalized to 1 because this corresponds to the resulting trade-offs for a single unit of change on the interface sliders. This trade-off matrix is constructed using the average swap distance between features across all permutations of column orderings and is generated using the algorithm outlined in Figure 2. Thus, as the user engages with the interface, the sliders adjust to the real world constraints of the problem space providing a tangible interactive representation, as shown below in Figure 3.

GenerateTradeOffMatrix(F, S, w) input: F set of n comparable features S set of m solutions w a function f(s) = w s = ranking decay function P set of all permutations of S W = empty n x n preference matrix for each p P fsum = empty vector of size n for each s S for each f F fsum[f] += s * w(s) fmap = empty map of size n for each f F fmap[f].put(f, fsum[f]) fmap.sort() for f 1 F for f 2 F if(f 1!= f 2) W[f 1, f 2] += fmap[f 1].pos - fmap[f 2].pos Figure 2 APPLYING STICKY SLIDERS TO THE COMPARISON INTERACTION While the application of weighted preference matrix R to the constrained slider interaction intuitively informs the user about the constraints of the problem space, the more desirable information is the user s own preference matrix P (defined similar to R) that allows a system to know how the individual user values different decision criteria. As expressed in prior research and literature, these preferences are often unknown prior, even to the user himself; however, these preferences are expressed latently in how the user reacts to the modified slider weights. The order in which a user responds to various changes in features (as well as the magnitude of that response i.e. the new level set by the user) indicates the particular trade-off values within the user s own matrix P. Figure 3

A critical assumption in the construction of the new comparison system is that replacing the generalized preference matrix R with the user s own ideal P will avoid the reactionary patterns of behavior and allow a user to naturally explore the item set according to his own preferences. Thus, a machine learning model is trained on the sequence of slide interaction data to identify points feature-to-feature reactions and construct a weighted matrix S such that: This modifying matrix P is returned to the client side interface upon each slide interaction and creates a sticking effect on the movement of feature sliders that keeps the explicit slider levels set by the user from moving with respect to the other values of R. LEARNING FROM USER ACTIONS WITH FEATURE-TO-FEATURE DECAY As the user corrects the weighted preference sliders, they generate a series of k actions, alternating between various features F 1, F 2,, F n for n features. This series of actions (called the actionlist) can be seen visually below, and the existing system keeps track of these state changes throughout the comparison session. In designing a useful training model, the primary design goal was that the features closest to each other within this actionlist are more strongly bound together (i.e. users are less willing to make tradeoffs between features F[i] and F[j] when C ij is small relative to the total size k of the actionlist). Conceptually, this corresponds to users reacting more strongly to a feature that was modified, indicating an expressed desire to correct the value to a previously set value. Thus, the final weighed update algorithm makes use of an exponential decay function of the distance from feature to feature within the action list, giving weighted preference to closer features. In calculating the user preferences S, a raw preference matrix S is initialized with empty values zero and is filled by performing a double loop through the action list and updating the matrix values: After this loop finishes, each row F[i] of S is normalized by calculating the row sum and dividing each row entry by this sum, where Δd i is the magnitude of change with each slider interaction. This generates a probability mass distribution of a user s expressed preferences regarding each F[i]-F[j] relationship; however, one must note that this preference matrix is not, in itself, of the same nature as the trade-off matrix R used to adjust the slide interactions. The final preference matrix S is derived:

In the live system, the final preference matrix S is used to return an updated trade-off matrix P, causing the user to adjust their interaction and comparison. (β = 0.25, τ = 1.0) DISCUSSION AND FUTURE WORK As presented here, the weighted decay approach to preference learning is uniquely situated within this limited domain of comparisons; however, the unique blend of machine learning and creative interaction design presents an entirely new research paradigm, especially within the realm of preference extraction. Typical approaches to latent variable discovery automatically assume the need for large datasets and complex unsupervised learning algorithms to discover hidden structure within user behavior patterns; however, this simple experiment and system shows that many of these problems can be alleviated with a more creative approach to interaction design. The online learning algorithm developed here iterates over the k items in the action list with each slider update (making an AJAX call to a back-end server); however, the calculations are not computationally complex for small numbers of features and solutions (which are constrained by human cognitive capabilities anyways). Future work will focus on using the preference information generated from these interactions as additional feature sources in traditional recommendation algorithms in an attempt to create a more robust query framework. The potential for using creative interaction patterns such as sticky sliders is incredible and presents a new paradigm in human computer interaction research. An incredible wealth of information lies in wait for the innovative interaction designer to begin capturing, and there is no telling what new improvements these data sources will lead to in both existing and future recommendation and information retrieval systems. REFERENCES Eastham, Garrett. (2010). HeyElroy: A Novel Approach to Feature Discovery in Information Retrieval Tasks Pu, P, Faltings, B, & Torrens, M. (2003). User-involved preference elicitation. IJCAI Workshop on Configuration Schyns, PG, Goldstone, RL, & Thibaut, JP. (1998). The Development of features in object concepts. Behavioral and Brain Sciences, 21(1), 1-54. Schyns, PG, & Rodet, L. (1997). Categorization creates functional features. Journal of experimental psychology. Learning, memory, and cognition, 23(3), 681. Slovic, P. (1995). The Construction of preference. American Psychologist, 50(5), 364-371.