Link Prediction Benchmarks

Similar documents
The link prediction problem for social networks

Online Social Networks and Media

Link Prediction and Anomoly Detection

SimRank : A Measure of Structural-Context Similarity

CS224W: Social and Information Network Analysis Project Report: Edge Detection in Review Networks

PageRank and related algorithms

CS 224W Final Report Group 37

Structural Analysis of Paper Citation and Co-Authorship Networks using Network Analysis Techniques

Topic mash II: assortativity, resilience, link prediction CS224W

Measuring scholarly impact: Methods and practice Link prediction with the linkpred tool

Using network evolution theory and singular value decomposition method to improve accuracy of link prediction in social networks

Social Information Retrieval

An Improved Computation of the PageRank Algorithm 1

DSCI 575: Advanced Machine Learning. PageRank Winter 2018

Centralities (4) By: Ralucca Gera, NPS. Excellence Through Knowledge

Collaborative Filtering using Euclidean Distance in Recommendation Engine

Fast Personalized PageRank On MapReduce Authors: Bahman Bahmani, Kaushik Chakrabart, Dong Xin In SIGMOD 2011 Presenter: Adams Wei Yu

Supervised Link Prediction with Path Scores

Modern Methods of Data Analysis - WS 07/08

Convexization in Markov Chain Monte Carlo

Equivalence Detection Using Parse-tree Normalization for Math Search

Semantic Annotation of Web Resources Using IdentityRank and Wikipedia

Web Spam Challenge 2008

Hierarchical Graph Clustering: Quality Metrics & Algorithms

Information Dissemination in Socially Aware Networks Under the Linear Threshold Model

Popularity of Twitter Accounts: PageRank on a Social Network

A Modified Algorithm to Handle Dangling Pages using Hypothetical Node

Roadmap. Roadmap. Ranking Web Pages. PageRank. Roadmap. Random Walks in Ranking Query Results in Semistructured Databases

Supplementary material to Epidemic spreading on complex networks with community structures

CS 229 Final Project - Using machine learning to enhance a collaborative filtering recommendation system for Yelp

Clustering: Classic Methods and Modern Views

Clustering. Robert M. Haralick. Computer Science, Graduate Center City University of New York

Uncertain Data Management Non-relational models: Graphs

Figure : Example Precedence Graph

Evaluation Metrics. (Classifiers) CS229 Section Anand Avati

Chapter 4: Implicit Error Detection

MAE 298, Lecture 9 April 30, Web search and decentralized search on small-worlds

Evaluation Measures. Sebastian Pölsterl. April 28, Computer Aided Medical Procedures Technische Universität München

Sentiment analysis under temporal shift

Link Prediction for Social Network

Personalized Document Rankings by Incorporating Trust Information From Social Network Data into Link-Based Measures

2013/2/12 EVOLVING GRAPH. Bahman Bahmani(Stanford) Ravi Kumar(Google) Mohammad Mahdian(Google) Eli Upfal(Brown) Yanzhao Yang

Degree Distribution: The case of Citation Networks

Collaborative filtering based on a random walk model on a graph

Web Spam. Seminar: Future Of Web Search. Know Your Neighbors: Web Spam Detection using the Web Topology

Link Analysis in Web Information Retrieval

Lecture 9: I: Web Retrieval II: Webology. Johan Bollen Old Dominion University Department of Computer Science

A brief history of Google

Samuel Coolidge, Dan Simon, Dennis Shasha, Technical Report NYU/CIMS/TR

10/10/13. Traditional database system. Information Retrieval. Information Retrieval. Information retrieval system? Information Retrieval Issues

Chapter 12: Query Processing. Chapter 12: Query Processing

QUINT: On Query-Specific Optimal Networks

Information Retrieval. Lecture 11 - Link analysis

Link Prediction in Graph Streams

An Empirical Analysis of Communities in Real-World Networks

Reddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011

Efficient query processing

Balancing Manual and Automatic Indexing for Retrieval of Paper Abstracts

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

Using Bloom Filters to Speed Up HITS-like Ranking Algorithms

α Coverage to Extend Network Lifetime on Wireless Sensor Networks

Web Structure Mining using Link Analysis Algorithms

Query Processing: The Basics. External Sorting

CS6200 Information Retrieval. David Smith College of Computer and Information Science Northeastern University

Lecture #3: PageRank Algorithm The Mathematics of Google Search

Exploring Econometric Model Selection Using Sensitivity Analysis

CLEF-IP 2009: Exploring Standard IR Techniques on Patent Retrieval

Supervised Random Walks on Homogeneous Projections of Heterogeneous Networks

Ranking Clustered Data with Pairwise Comparisons

CENTRALITIES. Carlo PICCARDI. DEIB - Department of Electronics, Information and Bioengineering Politecnico di Milano, Italy

Introduction to Parallel & Distributed Computing Parallel Graph Algorithms

Building Self-Service BI Solutions with Power Query. Written By: Devin

Parameterized Complexity of Independence and Domination on Geometric Graphs

2D rendering takes a photo of the 2D scene with a virtual camera that selects an axis aligned rectangle from the scene. The photograph is placed into

International Association of Scientific Innovation and Research (IASIR) (An Association Unifying the Sciences, Engineering, and Applied Research)

The PageRank Citation Ranking: Bringing Order to the Web

I/O Efficient Algorithms for Exact Distance Queries on Disk- Resident Dynamic Graphs

Tutorial 7: Automated Peak Picking in Skyline

Chapter 12: Query Processing

Encoding Words into String Vectors for Word Categorization

Entrance exam - Informatics

Inverted Indexes. Indexing and Searching, Modern Information Retrieval, Addison Wesley, 2010 p. 5

COMP Page Rank

Midterm Examination CS 540-2: Introduction to Artificial Intelligence

Capturing Missing Links in Social Networks Using Vertex Similarity

Cluster analysis. Agnieszka Nowak - Brzezinska

On Using Machine Learning for Logic BIST

Heap-Filter Merge Join: A new algorithm for joining medium-size relations

Semi-Supervised Clustering with Partial Background Information

Ranking web pages using machine learning approaches

A New Technique for Ranking Web Pages and Adwords

Recommendation System for Location-based Social Network CS224W Project Report

Towards a Wireless Lexicon. Philip Levis Computer Systems Lab Stanford University 20.viii.2007

CPSC 340: Machine Learning and Data Mining. Ranking Fall 2016

Categorical Data in a Designed Experiment Part 2: Sizing with a Binary Response

Analytical Techniques for Anomaly Detection Through Features, Signal-Noise Separation and Partial-Value Association

CPSC 532L Project Development and Axiomatization of a Ranking System

Ranking on Data Manifolds

Query Independent Scholarly Article Ranking

Retrieval of Web Documents Using a Fuzzy Hierarchical Clustering

Transcription:

Link Prediction Benchmarks Haifeng Qian IBM T. J. Watson Research Center Yorktown Heights, NY qianhaifeng@us.ibm.com October 13, 2016 This document describes two temporal link prediction benchmarks that can be downloaded at [10]. Data, code for evaluation and code for several predictors are included in the packages. Section 1 describes the first benchmark, which is link prediction in coauthorship social networks and is a replication of the experiments in [6]. Performances of several predictors are also presented. Section 2 describes the second benchmark, which is link prediction in Wikipedia. Performances of several predictors are also presented. Section 3 describes formats of data files and commands to run and evaluate predictors. Table 1: Statistics of the coauthorship networks. Entry format is our-number/number-reported-in- [6]. Column Collaborations denotes pairwise relations in the training period. Column E old denotes pairwise relations among Core authors in the training period. Column E new denotes new pairwise relations among Core authors formed in the test period. Training Period Core Authors Articles Collaborations Authors E old E new astro-ph 5321/5343 5820/5816 41575/41852 1563/1561 6189/6178 5719/5751 cond-mat 5797/5469 6698/6700 23373/19881 1387/1253 2367/1899 1514/1150 gr-qc 2150/2122 3292/3287 5800/5724 484/486 523/519 397/400 hep-ph 5482/5414 10277/10254 47788/47806 1786/1790 6629/6654 3281/3294 hep-th 5295/5241 9497/9498 15963/15842 1434/1438 2316/2311 1569/1576 1 Link prediction in arxiv The setup in [6] is the following. For five areas in arxiv, given the coauthors of papers published in the training period 1994-1996, the task is to predict new pairwise coauthorship relations formed in the test period 1997-1999. Predictions are only scored for those within Core authors, defined as those who have at least 3 papers in the training period and at least 3 papers in the test period; this Core list is unknown to predictors. Table 1 gives statistics of the five graphs and prediction tasks. Let E new be the set of new pairwise coauthorship relations among Core authors formed in the test period. Let E p be the top E new pairs among Core authors that are predicted by a predictor, and the score of this predictor is defined as E p E new / E new. Table 1 shows that our setup matches [6] closely for four of the five benchmarks. The download package include all five graphs. However, we will focus on benchmarks astro-ph, hep-ph and hep-th, for the following reasons. Benchmark cond-mat differs significantly from that reported in [6], thus is not a valid benchmark to compare against performances reported in [6]. In gr-qc, 131 out of the 397 new relations were formed by a single project which resulted in three papers in the test period, with nearly identical 45-46 authors, [1] being one of the three. Because the size of gr-qc is not large enough relative to this single event, the scores of the predictors are distorted. Thus it is not a surprise that [6] reported that the best predictor for gr-qc is one 1

Table 2: Comparison of predictor accuracies on coauthorship networks. A denotes the accuracy score of a predictor. R denotes the ratio of A over that of oracle of [6]. astro-ph hep-ph hep-th parameters A R A R A R Oracle of [6] varying 8.55% 7.2036% 7.9407% PPR, graph #2 α = 0.50 8.5330% 0.998 6.7358% 0.935 7.9031% 0.995 Modified Katz, graph #2 b 1 = 0.5, b 2 = 0.1, β = 0.1, γ = 5 8.4106% 0.984 8.2292% 1.142 7.8394% 0.987 ERD, graph #1 b 1 = 0.9, b 2 = 0.9 8.4281% 0.986 8.1682% 1.134 7.1383% 0.899 ERD, graph #2 b 1 = 0.9, b 2 = 0.9, γ = 4 9.5471% 1.117 8.7473% 1.214 7.1383% 0.899 that deletes 85-90% of edges as a preprocessing step, and that the same predictor delivers poor performance on the other benchmarks. Table 2 lists the accuracy of several predictors. Graph #1 refers to a uniformly weighted graph with edge (paper) weight b 1 and node (author) weight b 2. Graph #2 refers to the following graph. Each author is modeled as a node, with node weight: b author = 1 (1 b 2 ) 1/max(1,log γ m author) (1) where m author is the number of coauthors of this author in the training period. Each paper is modeled as a node, and it connects to and from each of its authors with two directed simple edges. An edge from node X to node Y has edge weight: b X,Y = 1 (1 b 1 ) 1/max(1,log γ dx) (2) where d X is the out degreeof node X. γ in (1)(2) is atunable parameterand is scannedwith other parameters and reported in Table 2. The listed predictors are: Oracle of [6] is the highest score reported for each benchmark, by all predictors including metaapproaches. The personalized PageRank (PPR) row uses [7, 8] and α is the teleportation probability. Since edge weights (2) are always equal for out-going edges from a same node, changing b 1 has no impact on transition probabilities and hence no impact on PPR scores. Since PageRank does not use node weights, we only need to scan parameter α as was done in [6]. The scores are asymmetric. We experimented with max, min, sum and product of two directional scores, and the max gives the best results and is reported in this row. The original Katz measure [5] does not use edge or node weights, and we define a modified Katz measure which does: score Katz (A,B) = β l (3) l=1 length-l A-to-B path i where β is a constant parameter, and p i,l is the product of edge weights and intermediate node weights for the i th path with length l. We experimented with max, min, sum and product of A-to-B and B-to-A scores, and the max gives the best results and is reported here. Expected Reliable Distance (ERD) is a distance measure from [9]. 10,000 Monte Carlo samples are used for each data point. For graph #2, the shortest-path distance in a Monte Carlo sample is defined as the shorter between the A-to-B path and the B-to-A path if both exist. We will further discuss its performance on hep-ph in Figure 1. Figure 1 shows detailed comparisons by receiver-operating-characteristic(roc) curves. The oracle of [6] is omitted because data is not available. Because the number of negative instances (pairs of authors) is far greater than the number of positive instances, the full ROC curves are not visually interesting. Therefore Figure 1 zooms in on the starting portions of curves where the true positive rate is less than 20% and the p i,l 2

(i) astro-ph (ii) hep-ph (iii) hep-th Figure 1: ROC curves on coauthorship networks. false positive rate is less than 1%; these portions are also the most meaningful in applications. Note that ERD #2 performs well on hep-ph for early guesses at around 5% true positive rate, but it degrades quickly after that and becomes the worst of the three by the 20% rate. 2 Link prediction in Wikipedia This section switches to concept domain: predicting additions of inter-wikipage citations in Wikipedia. The rationale is that citation links reflect Wikipedia contributors perception of relation strength between subjects. We obtain an inter-wikipage-citation graph from [3] which was based on Wikipedia dumps generated in April/May 2014, and another graph from [4] which was based on those in February/March 2015. In both graphs, each node represents a Wikipedia page, and each directed edge from node A to node B represents a citation on page A to page B. Statistics are shown in Table 3. 4,631,780 of the 2014 nodes are mapped to 2015 nodes by exact name matching, and another 87,368 are mapped to 2015 nodes by redirection data from [4] which are pages that have been renamed or merged. The remaining 12,396 of the 2014 nodes cannot be mapped: the majority are Wikipedia pages that have been deleted, and some are due to noise in the data collection of [3, 4]. Such noise is a small fraction and has negligible impact to our measurements. Table 3: Statistics of Wikipedia citation networks and prediction tasks. n page denotes the number of pages; d 2014 and d 2015 denote the average number of out-going citations on a 2014/2015 page; d 2014,unique and d 2015,unique denote the average number of unique out-going citations on a 2014/2015page; n prediction denotes the average number of additions to predict per task. n page d 2014 d 2014,unique d 2015 d 2015,unique n prediction 2014 all pages 4731544 24.66 23.99 2015 all pages 4964985 24.90 24.23 qualified tasks 93845 156.1 151.0 162.4 157.0 10.06 training tasks 1000 142.2 137.9 147.5 143.1 10.14 test tasks 1000 159.0 153.3 165.7 159.6 9.85 trimmed test tasks 949 157.6 151.7 164.3 157.9 4.63 3

For each mapped node A, we identify S A,2014 as the set of 2014 nodes that page A cites in the 2014 graph and that remain in the 2015 graph, S A,2015 as the set of 2014 nodes that page A cites in the 2015 graph, and X A,2014 as the set of 2014 nodes that cite page A. If page A satisfies the condition that 5 (S A,2015 \S A,2014 )\X A,2014 S A,2014 20%,weconsiderpageAasaqualifiedpredictiontask. Therationale behind the size limits is to choose test pages that have undergone thoughtful edits, and their 2014 page contents were already relatively mature; the rationale for excluding in-coming neighbors X A,2014 is to make the tasks more challenging, since simple techniques like heavily weighting in-coming edges have no effect. Statistics are shown in Table 3. The number of qualified tasks is large, and we randomly sample a 1000-task training set and a 1000-task test set. Mean average precision (MAP) [2] is the accuracy score. Nodes that do not exist in the 2015 graph or that exist in X A,2014 are removed from a predictor s outputs before MAP is measured. For each predictor, we choose its parameter(s) (α, β, b 1 and b 2 to be discussed later) to maximize MAP on the training set, and then under such parameterization, we report its MAP on the test set in Table 4 and Figure 2. Tasks vary in difficulty. If edits were to make a page more complete, the new links are often reasonably predictable and some are obvious. However, if edits were driven by a recent event, the new links are next to impossible to predict. We form a trimmed test set by removing from the test set targets that are too easy or too difficult. The removed prediction targets for a page A is added to X A,2014 so that they are excluded from predictor outputs before MAP evaluation. The results are listed in the last row of Table 3 and the last two columns of Table 4. Table 4: Comparison of predictor accuracies on additions of inter-wikipage citations in Wikipedia. Each predictor uses its best parameters selected based on the training set. R denotes the ratio of MAP of a predictor over MAP of Adamic/Adar. training test trimmed test parameters MAP R MAP R MAP R Adamic/Adar 0.0291 0.0281 0.0163 PPR, graph #1 α = 0.5 0.0299 1.029 0.0291 1.038 0.0186 1.140 PPR, graph #2 α = 0.2, γ = 500 0.0321 1.104 0.0309 1.100 0.0206 1.263 Modified Katz, graph #1 β = 5E-6 0.0269 0.925 0.0241 0.860 0.0151 0.924 Modified Katz, graph #2 b 1 = 0.8, b 2 = 0.8, β = 0.1, γ = 10 0.0341 1.173 0.0328 1.170 0.0198 1.213 ERD, graph #1 b 1 = 0.4, b 2 = 0.9 0.0266 0.914 0.0233 0.830 0.0162 0.996 ERD, graph #2 b 1 = 0.9, b 2 = 0.9, γ = 10 0.0238 0.817 0.0218 0.778 0.0154 0.944 Table 4 Lists accuracy of the various predictors. Graph #1 refers to a uniformly weighted graph with edge weight b 1 and node weight b 2. Graph #2 refers to a graph with the following weights. For an edge from node X to node Y and that represents the i th citation link on page X, its weight is The weight of a node is δ Y,X b edge = 1 (1 b 1 ) { 2, edge exists from Y to X where δ Y,X = 1, otherwise max(1,log γ i) max(1,log γ d Y,in) (4) b node = 1 (1 b 2 ) 1/(logd node,in+logd node,out ) The edge weight (4) gives higher weight to a citation link if it is located at an earlier location on a page, or if it points to a less-cited page, or if a returning citation exists. The node weight (6) is a direct adaptation of Adamic/Adar s (8). For best results, Katz and ERD use (4) as is while PPR uses a linear variant: b edge = b 1 δ Y,X max ( 1,log γ i ) max ( 1,log γ d Y,in ) (7) Unlike in Section 1, the relations to predict are asymmetric (page A adds a citation to page B) and hence there is no requirement for a score definition to be symmetric. Thus all predictors use their one-directional A-to-B score as is: (5) (6) 4

Adamic/Adar is implemented as: score Adamic/Adar (A,B) = C n A,C n C,B logd C,in +logd C,out (8) where n A,C is the number of A-to-C edges, n C,B is the number of C-to-B edges, and d C,in and d C,out are the numbers of in-coming and out-going edges of node C. When multiple node B s have the same score, we use their in-coming degrees as a tie breaker: a node with higher in-coming degree is ranked first. The same tie breaker is implemented in all other predictors, though its effect is minor to negligible. We use Adamic/Adar accuracy as a reference in Table 4, as it represents what can be achieved through good-quality local analysis. The personalized PageRank (PPR) rows use [7, 8] and α is the teleportation probability. PPR scores areinvariantunder different b 1 and b 2 values and we only need to find the α and γ values that maximize its MAP on the training set. Modified Katz uses (3). A remarkable observation about Modified Katz with graph #2 is that its best performance happens when (3) is almost divergent: with b 1 = 0.8, b 2 = 0.8 and γ = 10, the divergence limit for β is 0.1075. Expected Reliable Distance (ERD) is a distance measure from [9]. 100 Monte Carlo samples are used per task. The reduction in the number of samples, compared with Section 1, is because the Wikipedia graph is much larger and denser. Note that [9] used 50 samples in its experiments. (i) Test set (ii) Trimmed test set Figure 2: Accuracy curves on Wikipedia citation network. Figure 2 shows a more detailed comparison by plotting true positive rate as a function of the number of predictions made. Because the number of negative instances (pages not cited by a page) is far greater than the number of positives, the full curves are not visually interesting, and Figure 2 zooms in on the starting portions. 3 Data formats and command lines 3.1 arxiv data There are five folders. Each folder is one graph, and contains two files, kleinberg star graph.txt and kleinberg answer.txt. 5

The file kleinberg star graph.txt describes the topology of graph #2 in Section 1, and one can easily derive other forms of the coauthorship networks. The first line of kleinberg star graph.txt is: <number of nodes> <number of authors> where the number of nodes is equal to the number authors plus the number of papers. It is followed by <number of nodes> lines, each line describing the out-going edges of a node. Nodes have zero-based indexing. The first <number of authors> nodes are author nodes, while the rest are paper nodes. Each line has the following format: <out degree> <sink index> <sink index> <sink index> The file kleinberg answer.txt gives the correct answers of the prediction task. The fist line is <number of core authors> and is followed by <number of core authors> lines specifying each of the core authors indices. The next line is <number of new pairwise relations> and is followed by <number of new pairwise relations> lines, each in the following format: <author one> <author two> where the first index is always the smaller of the two. There are five.cpp files in the download package. The four predictor *.cpp files are source code of the four predictors in Table 2. The check answers.cppfile is source code to check the output of predictors. They each can be compiled into executables simply by The command lines to run the four predictors are g++ -O3 foo.cpp -o foo predictor pagerank 2 -data kleinberg star graph.txt -output myanswers -alpha 0.5 predictor katz 2 -data kleinberg star graph.txt -output myanswers -beta 0.1 -weight1 0.5 -weight2 0.1 predictor relidist 1 -data kleinberg star graph.txt -output myanswers -weight1 0.9 -weight2 0.9 predictor relidist 2 -data kleinberg star graph.txt -output myanswers -weight1 0.9 -weight2 0.9 The command line to evaluate the output of a predictor is check answers -input myanswers -correctanswers kleinberg answer.txt -report my.report -roc my.roc which produces two text files my.report and my.roc, the first reporting the accuracy score shown in Table 2 and the second reporting the ROC curve in Figure 1. 3.2 Wikipedia data The file page links en.edges describes the graph, and files page links en.bidi, page links en.names and page links en.gone.index provide supplemental information. Files queries.training and answers.training are the 1000 training tasks and their correct answers. Files queries.test and answers.test are the 1000 test tasks and answers. Files queries.trimmed test, answers.trimmed test and excluded answers.trimmed test are the trimmed test tasks and answers. In file page links en.edges, nodes have zero-based indexing, and each line describes the out-going edges of a node and has the format: <out degree> <sink index> <sink index> <sink index> 6

In file page links en.bidi, each line corresponds to one line in page links en.edges, and each 0/1 specifies whether each out-going edge has a return citation. Obviously this file is not necessary and can be derived from page links en.edges on the fly. The purpose here is to simplify predictor code. The file page links en.gone.index lists indices of nodes that do not exist in 2015 Wikipedia. Therefore these nodes are removed from output predictions of a predictor before MAP is calculated. This is only used for evaluation purpose and a predictor shall not have access to this information. The file page links en.names maps indices to Wikipedia page titles, for your amusement. Each file queries.* lists the starting points of prediction tasks. Each file answers.* has the following format per line. <start index> <number of answers> <answer index> <answer index> <answer index> The trimmed test has an extra file excluded answers.trimmed test, which has the following format per line. <number of entries> <index> <index> <index> These are answers that have been trimmed from the test, and therefore during evaluation these are removed from output predictions of a predictor before MAP is calculated. There are eight.cpp files in the download package. The seven predictor *.cpp files are source code of the seven predictors in Table 4. The check answers.cpp file is source code to check the output of predictors. Different from the arxiv data, if we let each predictor write out all its predictions for 1000 tasks, they would take a large amount of disk space. Therefore, we embed evaluation code into the prediction runs by including check answers.cpp in predictor *.cpp files. The predictor *.cpp files each can be compiled into executables simply by g++ -O3 foo.cpp -o foo With certain gcc versions, extra includes are needed. For example, you may need to add a line #include <algorithm> to check answers.cpp. The command lines to run the seven predictors, and to evaluate their outputs in the same run, are predictor adamicadar -data page links en.edges -query queries.<x> -correctanswers answers.<x> \\ -ignore page links en.gone.index -report my.report -curve my.plot predictor pagerank 1 -data page links en.edges -query queries.<x> -correctanswers answers.<x> \\ -ignore page links en.gone.index -report my.report -curve my.plot \\ -alpha 0.5 predictor pagerank 2 -data page links en.edges -query queries.<x> -correctanswers answers.<x> \\ -ignore page links en.gone.index -report my.report -curve my.plot \\ -bidi page links en.bidi -alpha 0.2 predictor katz 1 -data page links en.edges -query queries.<x> -correctanswers answers.<x> \\ -ignore page links en.gone.index -report my.report -curve my.plot predictor katz 2 -data page links en.edges -query queries.<x> -correctanswers answers.<x> \\ -ignore page links en.gone.index -report my.report -curve my.plot \\ -bidi page links en.bidi predictor relidist 1 -data page links en.edges -query queries.<x> -correctanswers answers.<x> \\ -ignore page links en.gone.index -report my.report -curve my.plot predictor relidist 2 -data page links en.edges -query queries.<x> -correctanswers answers.<x> \\ -ignore page links en.gone.index -report my.report -curve my.plot \\ -bidi page links en.bidi If a run is on the trimmed test set, an extra argument is needed: -exclude excluded answers.trimmed test. Arun produces twotext filesmy.reportandmy.plot,the first reportingthe accuracyscoreshownin Table4 and the second reporting the curve in Figure 2. 7

References [1] A. M. Abrahams et al. Gravitational wave extraction and outer boundary conditions by perturbative matching. arxiv:gr-qc/9709082, 1997. [2] R. Baeza-Yates and B. Ribeiro-Neto. Modern Information Retrieval. Addison Wesley, 1999. [3] DBpedia. DBpedia 2014 Downloads, 2014. [Available at http://wiki.dbpedia.org/downloads2014; accessed 22-April-2015]. [4] DBpedia. DBpedia Downloads 2015-04, 2015. [Available at http://wiki.dbpedia.org/downloads2015-04; accessed 20-August-2015]. [5] L. Katz. A new status index derived from sociometric analysis. Psychometrika, 18(1):39 43, 1953. [6] D. Liben-Nowell and J. Kleinberg. The link-prediction problem for social networks. Journal of the American Society for Information Science and Technology, 58(7):1019 1031, 2007. [7] L. Page. Method for node ranking in a linked database. U.S. Patent 6285999, 2001. [8] L. Page, S. Brin, R. Motwani, and T. Winograd. The pagerank citation ranking: bringing order to the web. Technical Report, Stanford University, 1999. [9] M. Potamias, F. Bonchi, A. Gionis, and G. Kollios. K-nearest neighbors in uncertain graphs. Proceedings of the VLDB Endowment, 3(1):997 1008, 2010. [10] H. Qian. Link prediction benchmarks, 2016. [Available at http://researcher.watson.ibm.com/group/6672]. 8