HUMAN ACCURACY ANAYLSIS ON THE AMAZON MECHANICAL TURK

Similar documents
MS&E 226: Small Data

Using Machine Learning to Optimize Storage Systems

Fall 09, Homework 5

Regularization and model selection

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov

Clustering with Reinforcement Learning

Character Recognition

Lab 9. Julia Janicki. Introduction

Machine Learning Classifiers and Boosting

Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions

CS 188: Artificial Intelligence Fall Machine Learning

10-701/15-781, Fall 2006, Final

Gradient Descent. Wed Sept 20th, James McInenrey Adapted from slides by Francisco J. R. Ruiz

1 Machine Learning System Design

CS231A Course Project Final Report Sign Language Recognition with Unsupervised Feature Learning

Mixture Models and the EM Algorithm

MIT Samberg Center Cambridge, MA, USA. May 30 th June 2 nd, by C. Rea, R.S. Granetz MIT Plasma Science and Fusion Center, Cambridge, MA, USA

To earn the extra credit, one of the following has to hold true. Please circle and sign.

Unsupervised Learning

CS6220: DATA MINING TECHNIQUES

Ensemble methods in machine learning. Example. Neural networks. Neural networks

COMP 551 Applied Machine Learning Lecture 13: Unsupervised learning

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas

4 KEY FACTORS FOR DATA QUALITY ON A DATA LAKE (OR: HOW TO AVOID THE DATA SWAMP) JOSH HERRITZ MIOSOFT CORPORATION MIOsoft Corporation.

Bayesian Methods in Vision: MAP Estimation, MRFs, Optimization

Alternatives to Direct Supervision

5 Learning hypothesis classes (16 points)

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

Lecture 3: Linear Classification

CS6375: Machine Learning Gautam Kunapuli. Mid-Term Review

Network Traffic Measurements and Analysis

GAMs semi-parametric GLMs. Simon Wood Mathematical Sciences, University of Bath, U.K.

Paul's Online Math Notes Calculus III (Notes) / Applications of Partial Derivatives / Lagrange Multipliers Problems][Assignment Problems]

Lecture 1 Notes. Outline. Machine Learning. What is it? Instructors: Parth Shah, Riju Pahwa

Rapid Software Testing Guide to Making Good Bug Reports

CSE 573: Artificial Intelligence Autumn 2010

CITS4009 Introduction to Data Science

CLUSTERING. JELENA JOVANOVIĆ Web:

Final Exam. Introduction to Artificial Intelligence. CS 188 Spring 2010 INSTRUCTIONS. You have 3 hours.

The Expectation Maximization (EM) Algorithm

Lecture 19: Generative Adversarial Networks

Week 3: Perceptron and Multi-layer Perceptron

Based on Raymond J. Mooney s slides

Dynamic Routing Between Capsules

Dynamic Thresholding for Image Analysis

CS 540: Introduction to Artificial Intelligence

Predicting User Ratings Using Status Models on Amazon.com

Part-Based Skew Estimation for Mathematical Expressions

Lecture 25: Review I

COMPUTATIONAL INTELLIGENCE SEW (INTRODUCTION TO MACHINE LEARNING) SS18. Lecture 6: k-nn Cross-validation Regularization

Unsupervised learning in Vision

Bayes Net Learning. EECS 474 Fall 2016

Analytical model A structure and process for analyzing a dataset. For example, a decision tree is a model for the classification of a dataset.

Semi-Supervised Clustering with Partial Background Information

Louis Fourrier Fabien Gaie Thomas Rolf

CPSC 340: Machine Learning and Data Mining. Probabilistic Classification Fall 2017

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

CSC 258 lab notes, Fall 2003

Today. Lecture 4: Last time. The EM algorithm. We examine clustering in a little more detail; we went over it a somewhat quickly last time

Elemental Set Methods. David Banks Duke University

CS 188: Artificial Intelligence Fall Announcements

Announcements. CS 188: Artificial Intelligence Fall Machine Learning. Classification. Classification. Bayes Nets for Classification

Clustering K-means. Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, Carlos Guestrin

Lessons Learned from Large Scale Crowdsourced Data Collection for ILSVRC. Jonathan Krause

Chapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea

Multi-label classification using rule-based classifier systems

MIT 801. Machine Learning I. [Presented by Anna Bosman] 16 February 2018

The Mathematics Behind Neural Networks

Notes on Multilayer, Feedforward Neural Networks

The exam is closed book, closed notes except your one-page (two-sided) cheat sheet.

Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah

A noninformative Bayesian approach to small area estimation

Online Algorithm Comparison points

Announcements. CS 188: Artificial Intelligence Spring Classification: Feature Vectors. Classification: Weights. Learning: Binary Perceptron

Singular Value Decomposition, and Application to Recommender Systems

Evaluation. Evaluate what? For really large amounts of data... A: Use a validation set.

Crowdsourcing Annotations for Visual Object Detection

Pull the Plug? Predicting If Computers or Humans Should Segment Images Supplementary Material

CS 188: Artificial Intelligence Fall 2011

Tutorial: Using Tina Vision s Quantitative Pattern Recognition Tool.

The Fly & Anti-Fly Missile

9 th CA 2E/CA Plex Worldwide Developer Conference 1

Machine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme, Nicolas Schilling

A Formal Approach to Score Normalization for Meta-search

Automated Crystal Structure Identification from X-ray Diffraction Patterns

A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models

Feature Extractors. CS 188: Artificial Intelligence Fall Some (Vague) Biology. The Binary Perceptron. Binary Decision Rule.

Classification: Feature Vectors

A Comparison of Error Metrics for Learning Model Parameters in Bayesian Knowledge Tracing

Chapter 2 Basic Structure of High-Dimensional Spaces

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

In the recent past, the World Wide Web has been witnessing an. explosive growth. All the leading web search engines, namely, Google,

Big Data Methods. Chapter 5: Machine learning. Big Data Methods, Chapter 5, Slide 1

Unsupervised Outlier Detection and Semi-Supervised Learning

Rapid Natural Scene Text Segmentation

Statistics Case Study 2000 M. J. Clancy and M. C. Linn

Probabilistic Abstraction Lattices: A Computationally Efficient Model for Conditional Probability Estimation

Music Genre Classification

Problem 1 (20 pt) Answer the following questions, and provide an explanation for each question.

CP365 Artificial Intelligence

Transcription:

HUMAN ACCURACY ANAYLSIS ON THE AMAZON MECHANICAL TURK JASON CHEN, JUSTIN HSU, STEFAN WAGER Platforms such as the Amazon Mechanical Turk (AMT) make it easy and cheap to gather human input for machine learning problems. The flipside, however, is that some human contributors can be confused or inaccurate. The challenge then becomes to extract trustworthy information from the answers of more or less accurate contributors. Image-Net is an effort to organize images collected online into hierarchal synonym sets (synsets), such as dog or jigsaw puzzle. Image-Net then employs AMT turkers to label whether the images it has gathered are actually typical representatives of the synset they were grouped under. Since the turker labels are not 100% accurate, Image-Net currently determines whether an image is correctly labeled by majority vote. In this project, we consider machine learning algorithms that could enable us to identify the more accurate turkers, and to recognize more typical representatives of synsets using the available AMT votes. First, we present an unsupervised EM algorithm that assigns accuracies to individual turkers and assumes that their votes are Bernoulli-distributed. Second, we present a supervised algorithm that identifies a few dozen key images, which we correctly label, and then uses this information to identify trustworthy turkers. The supervised approach produced some very promising results, especially in synsets with ambiguous definitions, such as bishop (the chess piece); the unsupervised results were a bit more mixed. In both our approaches, we make the following assumptions: first, we assume that each of our voters has a consistent accuracy level; second, we assume that the majority is more often correct than not. We then attempt to identify turkers who vote with the majority most often, infer that they are our best labelers, and follow these trusted labellers in ambiguous cases. Unsupervised Learning Approaches First Approaches: Naive EM and Skill Scoring. Before settling on our final design, we attempted a Naive EM approach. Briefly, we assume each turker has a weight c i corresponding to their accuracy. We set the weights to be the log odds of the fraction of the turker s votes that are correct, then we adjusted the labels of the pictures based on the weighted majority vote. Though this seemed promising at first, we encountered one of the main difficulties of this project: too many negative images. Since most of the images the turkers are presented with are negative examples, a turker who votes No to all images will score very well. Thus, these people will have disproportionately high weights, skewing our labels, and leading to mediocre performance. To remedy this, we considered using a skill scoring approach, in which we roughly estimate the fraction of positive images. With this number, we can calculate baseline performances of turkers, if they simply always vote Yes or always vote No. By taking this baseline performance into account, we can decrease the bias. The main problem with this approach is that it isn t clear how to normalize the performances. A turker s maximum performance depends not only on how many images she votes on, but also on the votes she casts, as it is easier to get a No vote correct than a Yes vote. After trying several normalizing schemes, we found that the results were much more promising than the first approach, with poor performance only in identifying false positives. Final Approach: Specificity and Sensitivity EM. We went back to our first approach, and looked for a model in which voters were expected to score correctly at a high rate when voting No, and to correct for Date: November 2010. 1

2 JASON CHEN, JUSTIN HSU, STEFAN WAGER this bias. Our idea was to have different weights for when turkers voted Yes, and for when turkers voted No. This led to the following model (derivation closely follows [1]): Let y = 0, 1 be the ground truth of a given image, and y j = 0, 1 is voter j labeled the image as. For voter j, we associate parameters α j = Pr[y j = 1 y = 1], β j = Pr[y j = 0 y = 0], known as the sensitivity and the specificity, respectively. We also have one more hidden parameter, p = Pr[y = 1], or the fraction of typical images amongst all images. Now, if D is the observed votes, y uj i is the vote given to picture i by turker u j, the jth turker to vote on this picture, and θ are the parameters α j, β j, p, we have maximum likelihood function: Pr[D θ] = N Pr[y 1 i,..., y Ri i θ] = Independence of votes gives: N { a i := Pr[y 1 i,..., y Ri i y i = 1, α] = b i := Pr[y 1 i,..., y Ri i y i = 0, β] = Pr[y 1 i,..., y Ri i } y i = 1, α] p + Pr[yi 1,..., y Ri i y i = 0, β] (1 p) R i Pr[y j i y R i i = 1, α j ] = (α j ) yj i (1 α j ) 1 yj i R i Pr[y j i y R i i = 0, β j ] = (β j ) 1 yj i (1 β j ) yj i Thus we want to maximize Pr[D θ] = N a ip + b i (1 p). Denoting the ground truth as y = [y 1,..., y N ], we get that the log-likelihood is N y i ln p i a i + (1 y i ) ln(1 p i )b i E-step: We calculate the expected value of this log-likelihood (this will be our lower bound for the true log-likelihood). Taking the expectation with respect to Pr[y D, θ] = Q i (y), we have (1) E[ln Pr[D, y θ]] = N µ i ln p i a i + (1 µ i ) ln(1 p i )b i where µ i = Pr[y i = 1 yi 1,..., yri i, θ]. By Bayes, we update µ i Pr[y 1 i,..., y Ri i y i = 1, θ] Pr[y i = 1] µ i := a i p a i p + b i (1 p) M-step: We wish to maximize the lower bound calculated in the E-step with respect to the parameters α j, β j. By setting the α and β gradients of (1) to 0 and update p, we get: N α k := µ iyi k N N µ β k := (1 µ i)(1 yi k ) N i (1 µ p := 1 N µ i i) N By reestimating p = 1 N N µ i every iteration, we can run this until convergence, initializing µ i to be the majority vote µ i = Pr[y i = 1 yi 1,..., yri i ] = 1 Ri N yj i To calculate our final labels, since µ i is the probability that image i is typical, we can set a threshold γ and mark images that have µ i > γ as typical. Results. The results from our unserpervised approach were somewhat mixed. For some synsets, we managed to perform better than majority vote; on others, though, we generated a large number of false positives. Below are results for the following three synsets, as well as their definitions according to Image-Net: Caterpillar, Cat: a large tracked vehicle that is propelled by two endless metal belts; frequently used for moving earth in construction and farm work. Bishop: (chess) a piece that can be moved diagonally over unoccupied squares of the same color. Horseshoe: U-shaped plate nailed to underside of horse s hoof. For each synset, we record the number of positives the algorithm recognized, the fraction of images that were unambiguously typical, and the false positive rate. We did not count ambiguous images in either of these categories.

HUMAN ACCURACY ANAYLSIS ON THE AMAZON MECHANICAL TURK 3 Figure 1. Sensitivity (left) and specificity (right) of users in the Caterpillar, Cat synset Results Images Recognized Definitely Typical False Positives Cat-Cat: Unsupervised 531 66% 13% Cat-Cat: Majority Vote 579 45% 39% Bishop: Unsupervised 280 52% 43 % Bishop: Majority Vote 167 91% 3% Horseshoe: Unsupervised 1422 75% 16% Horseshoe: Majority Vote 1248 72% 13% Our algorithm worked fairly well for Caterpillar, Cat. However, it performed poorly for Bishop. We suspect this is because a very small fraction of the images presented to the AMT are in fact typical: under 10%. This means that voting on no-images is very easy, while getting a high score on yes-images is hard. Thus, most turkers have very high specificities but low sensitivities. Such numbers tend to push split votes toward a yes-decision, since our high specificity and low sensitivity reflect an assumption by the model that false positives are quite unlikely, whereas false negatives are common. If we try to fix this problem by computing only a single credibility for each turker (as we did in our first model), we run into another related problem: all the easy no-votes give most turkers an artificially high credibility. A challenge for unsupervised learning approaches to the Image-Net dataset is that the interesting images are swamped by a large mass of easy no-images from which it is difficult to learn anything new. In the next section, we present an alternative approach that only learns turker accuracies from contentious, or difficult images. Supervised Learning Approach We suspect our work unsupervised learning algorithms ran into difficulties for the following reasons: Learning contributor accuracies from all images is not very effective, because interesting, difficult images are swamped by easy images on which everyone votes correctly. Learning contributor accuracies from the most difficult images in an unsupervised way can be dangerous, since even a large majority can be wrong on difficult images. One way to overcome these problems is to automatically find the most difficult and contentious images, and then have an accredited expert provide the ground truth on these images. Though this seems to defeat the purpose of using AMT input, we think that this approach is valuable if a researcher really needs accurate labels on a set of tens of thousands of images and has time to manually label a few dozen images per synset to train the supervised learning algorithm. After these labels are set, the full collection of AMT votes can then be used effectively.

4 JASON CHEN, JUSTIN HSU, STEFAN WAGER Our idea is an example of a collaborative filtering approach, in that we estimate of how each turker might have voted on our small labeled training set by comparing them to turkers that actually voted on our training set. We then compute weights for each turker using a Bernoulli maximum likelihood estimate, and infer labels for our images using a weighted majority vote. Specifically, we first pick 16 images with a negative majority vote and 16 images with a positive majority vote, while maximizing the quantity (#yes-votes #no-votes) for each image. These are our contentious images form our key images, which we label manually: for each key image k l we provide a definite label L(k l ) = 0, 1. To compute weights for all turkers, we first need to to obtain predictions Ãik l for how each turker i would have voted on key image k l. Clearly, if i voted on k l, then Ãik l = A ikl. However, if i did not vote on k l, we need to somehow guess how i would have voted on k l. To do this, we first compute the similarity coefficient u ij for all turkers i and j: u ij = k images A ika jk k images A ika jk Assuming that turkers {l j } voted on key image k l, we use these parameters to give a guess for the vote of i / {l j } on k l : Ã ikl = j u il j A ljk l. We are now ready to compute weights for each turker. Assuming that the votes by turker i are Bernoulli-distributed such that they are accurate with probability c i, we get the following experssion to maximize: c max(l(kl)aikl,0) i,k l i (1 c i ) max( L(k l)a ikl,0) With a little math (see milestone), we get that c i = P k l L(k l )A ikl P k l L(k l )A ikl. Taking the log of the likelihood function, we can also find additive weights w i = log( ci 1 c i ). We do not let additive weights drop below zero; instead, we ignore turkers with negative weights. Having computed additive weights for each turker as described, we proceed to a weighted majority vote with which we infer labels for the remaining images. Results. The results using our supervised algorithm were quite promising. Especially for very confusing synsets, such as Caterpillar, Cat, our algorithm did a good job at weeding out turkers working under a wrong definition. Below are some detailed results; the Yes and No on the second row indicate the decisions by our algorithm. Majority Vote No Majority Vote Yes Caterpillar Cat No Yes No Yes Total Number of Images 12232 159 297 282 Sample Size 100 100 100 100 Number of Definite Positive 2 55 8 78 Number of Ambiguous 3 38 13 20 Number of Definite Negative 95 7 79 2 Many images, such as vehicles with wheels manufactured by Cat, or vehicles with tracks but not directly useable for farming, were ambiguous as to whether they should be considered to represent Caterpillar, Cat or not; we labeled these ambiguous. The sample size row indicates how many images we looked at in each category to evaluate our algorithm. Evidently, in all the cases where our algorithm disagreed with the majority vote, it had less than 1 chance in 10 of making a non-ambiguous mistake. From the point of view of Image-Net, which is trying to accumulate a large and accurate database of images, the two most important metrics are the number of positive images recognized, and the false positive rate. We provide such statistics below for the 3 synsets we worked with:

HUMAN ACCURACY ANAYLSIS ON THE AMAZON MECHANICAL TURK 5 Results Images Recognized Definitely Typical False Positives Cat-Cat: Supervised 464 70% 4% Cat-Cat: Majority Vote 579 45% 39% Bishop: Supervised 190 89% 0% Bishop: Majority Vote 167 91% 3% Horseshoe: Supervised 1259 73% 12% Horseshoe: Majority Vote 1248 72% 13% Our algorithm performed quite well on both Caterpillar, Cat and Bishop. We notice that both of these synsets have somewhat confusing definitions: some might mistakenly believe that the first applies to larvae of butterflies, and the second to Catholic clergy. Our hypothesis is that our algorithm managed to eliminate confused voters, and thus clean up the vote dataset. Our algorithm does not, however, perform nearly as well on Horseshoe. The difficulty in picking good images of horseshoes is not so much due to confusing definitions, but rather in learning what to do with limit cases such as horseshoe shaped necklaces. These results suggest that our algorithm is good at picking out confused turkers, but not so good at picking out turkers without a keen attention to detail. As support for the hypothesis that our algorithm works essentially by eliminating confused labellers, consider the following numbers: With Caterpillar, Cat, we gave 20 out of 80 turkers a weight of 0, and with Bishop, 10 out of 28 voters got no weight. On the other hand, with Horseshoe, we only gave 4 out of 156 turkers a weight of 0. This suggests that we did poorly on the latter synset because we couldn t identify the turkers who were consistently wrong on a majority of the key images. Discussion and Further Works Our results show that machine learning techinques can successfully be applied to improve the value of votes gathered on AMT. Our supervised approach, in particular, makes it easy for a researcher who is willing to spend about a minute labelling images from a synset to eliminate data from turkers who do not share her understanding definitions of the synsets. Our unsupervised approach presents some interesting challenges. The central difficulty is that, for many synsets, the number of bad images swamps the number of interesting images. If we give turkers a simple Bernoulli accuracy, the high number of trivially bad images makes this accuracy artificially high. Conversely, if we try to give each turker a separate sensitivity and specificity, we end up giving each turker an extremely high specificity. This results in unwanted false positives, since having very high specificities makes it unlikely that any image having received many yes-votes could be a negative. The success of the supervised approach suggests that it is most useful to infer turker accuracies from their performance on the most difficult images. An interesting problem for further study would be to work on unsupervised learning algorithms that focus more closely on contentious images. One way this could be implemented is by ignoring all images with unanimous votes while learning accuracies. Another approach could be to introduce a parameter for the difficulty of each image the probability of a turker making a mistake could then depend both on her accuracy and on the difficulty of the image. See, for example, [2] for an idea similar to the latter. Our results show that there is much promise and much room for improvement in automatic AMT vote analysis. Since AMT is one of the most promising ways of collecting machine learning data on a large scale, it is critical that we find good ways of analyzing AMT data as efficiently as possible. We look forward to seeing what approaches will surface in the literature over the next years. References [1] V. Rayker, et al. Learning From Crowds, Journal of Machine Learning Research, 11 (2010), 1297 1322 [2] J. Whitehill, et al. Whose Vote Should Count More: Optimal Integration of Labels from Labelers of Unknown Expertise, Advances in Neural Information Processing Systems (forthcoming), 2009.