Bias-free Hypothesis Evaluation in Multirelational Domains

Similar documents
A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection (Kohavi, 1995)

Learning Relational Probability Trees Jennifer Neville David Jensen Lisa Friedland Michael Hay

Combining Gradient Boosting Machines with Collective Inference to Predict Continuous Values

Estimating Missing Attribute Values Using Dynamically-Ordered Attribute Trees

Evaluating the Replicability of Significance Tests for Comparing Learning Algorithms

A Shrinkage Approach for Modeling Non-Stationary Relational Autocorrelation

Evaluation Strategies for Network Classification

Relational Learning. Jan Struyf and Hendrik Blockeel

ECLT 5810 Evaluation of Classification Quality

Sample 1. Dataset Distribution F Sample 2. Real world Distribution F. Sample k

Stochastic propositionalization of relational data using aggregates

Multi-label classification using rule-based classifier systems

Linkage and Autocorrelation Cause Feature Selection Bias in Relational Learning

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

A Two-level Learning Method for Generalized Multi-instance Problems

What is Learning? CS 343: Artificial Intelligence Machine Learning. Raymond J. Mooney. Problem Solving / Planning / Control.

Collective Classification with Relational Dependency Networks

Link Prediction for Social Network

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

An Information-Theoretic Approach to the Prepruning of Classification Rules

Challenges motivating deep learning. Sargur N. Srihari

Mining di Dati Web. Lezione 3 - Clustering and Classification

Application of Support Vector Machine Algorithm in Spam Filtering

Performance Estimation and Regularization. Kasthuri Kannan, PhD. Machine Learning, Spring 2018

Salford Systems Predictive Modeler Unsupervised Learning. Salford Systems

E-Companion: On Styles in Product Design: An Analysis of US. Design Patents

Sampling PCA, enhancing recovered missing values in large scale matrices. Luis Gabriel De Alba Rivera 80555S

Using Decision Boundary to Analyze Classifiers

A Model of Machine Learning Based on User Preference of Attributes

Nonparametric Methods Recap

Probabilistic Abstraction Lattices: A Computationally Efficient Model for Conditional Probability Estimation

Machine Learning Techniques for Data Mining

Semi-Supervised Clustering with Partial Background Information

Mining High Order Decision Rules

An Empirical Study of Hoeffding Racing for Model Selection in k-nearest Neighbor Classification

Slides for Data Mining by I. H. Witten and E. Frank

A Comparison of Text-Categorization Methods applied to N-Gram Frequency Statistics

Joint Entity Resolution

CPSC 340: Machine Learning and Data Mining. Probabilistic Classification Fall 2017

MULTI-VIEW TARGET CLASSIFICATION IN SYNTHETIC APERTURE SONAR IMAGERY

Relational Classification for Personalized Tag Recommendation

Graph Classification in Heterogeneous

WEKA: Practical Machine Learning Tools and Techniques in Java. Seminar A.I. Tools WS 2006/07 Rossen Dimov

Knowledge Discovery and Data Mining

Efficient Pairwise Classification

Core Membership Computation for Succinct Representations of Coalitional Games

Artificial Intelligence. Programming Styles

Efficient Case Based Feature Construction

Processing Missing Values with Self-Organized Maps

Maximum number of edges in claw-free graphs whose maximum degree and matching number are bounded

SSV Criterion Based Discretization for Naive Bayes Classifiers

Section 4 Matching Estimator

Hybrid Feature Selection for Modeling Intrusion Detection Systems

City, University of London Institutional Repository

Relational Classification Through Three State Epidemic Dynamics

Using Machine Learning to Optimize Storage Systems

Machine Learning (CSE 446): Concepts & the i.i.d. Supervised Learning Paradigm

Unsupervised Learning and Clustering

Multi-relational Decision Tree Induction

Learning Directed Probabilistic Logical Models using Ordering-search

Improving Imputation Accuracy in Ordinal Data Using Classification

Clustering Part 4 DBSCAN

SIMILARITY MEASURES FOR MULTI-VALUED ATTRIBUTES FOR DATABASE CLUSTERING

Random Forest A. Fornaser

Machine Learning Techniques for Data Mining

Evaluating Machine-Learning Methods. Goals for the lecture

A noninformative Bayesian approach to small area estimation

A Well-Behaved Algorithm for Simulating Dependence Structures of Bayesian Networks

Collective classification in network data

Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic

Generative and discriminative classification techniques

Data Mining. 3.3 Rule-Based Classification. Fall Instructor: Dr. Masoud Yaghini. Rule-Based Classification

Chapter 4 Fuzzy Logic

Outlier Ensembles. Charu C. Aggarwal IBM T J Watson Research Center Yorktown, NY Keynote, Outlier Detection and Description Workshop, 2013

Evaluating Classifiers

Chapter 3: Supervised Learning

Performance Analysis of Data Mining Classification Techniques

Adaptive Metric Nearest Neighbor Classification

CHAPTER 4: CLUSTER ANALYSIS

Character Recognition

Stability of Feature Selection Algorithms

Workload Characterization Techniques

Improving Quality of Products in Hard Drive Manufacturing by Decision Tree Technique

Recent Progress on RAIL: Automating Clustering and Comparison of Different Road Classification Techniques on High Resolution Remotely Sensed Imagery

Indexing by Shape of Image Databases Based on Extended Grid Files

Efficient SQL-Querying Method for Data Mining in Large Data Bases

University of Florida CISE department Gator Engineering. Clustering Part 4

The strong chromatic number of a graph

Enhancing Cluster Quality by Using User Browsing Time

Feature-weighted k-nearest Neighbor Classifier

Efficient Multi-relational Classification by Tuple ID Propagation

The Role of Biomedical Dataset in Classification

A Survey on Postive and Unlabelled Learning

CSC411/2515 Tutorial: K-NN and Decision Tree

K- Nearest Neighbors(KNN) And Predictive Accuracy

Byzantine Consensus in Directed Graphs

Detecting Data Structures from Traces

Improving Naïve Bayes Classifier for Software Architecture Reconstruction

Dependency Networks for Relational Data

A Unified Framework to Integrate Supervision and Metric Learning into Clustering

Transcription:

Bias-free Hypothesis Evaluation in Multirelational Domains Christine Körner Fraunhofer Institut AIS, Germany christine.koerner@ais.fraunhofer.de Stefan Wrobel Fraunhofer Institut AIS and Dept. of Computer Science III University of Bonn, Germany stefan.wrobel@ais.fraunhofer.de Abstract In propositional domains, using a separate test set via random sampling or cross validation is generally considered to be an unbiased estimator of true error. In multirelational domains, previous work has already noted that linkage of objects may cause these procedures to be biased, and has proposed corrected sampling procedures. However, as we show in this paper, the existing procedures only address one particular case of bias introduced by linkage. We recall that in the propositional case cross validation measures off-training set OTS error and not true error and illustrate the difference with a small experiment. In the multirelational case, we show that the distinction between training and test set needs to be carefully extended based on a graph of potentially linked objects, and on their assumed probabilities of reoccurrence. We demonstrate that the bias due to linkage to known objects varies with the chosen proportion of the training/test split and present an algorithm, generalized subgraph sampling, that is guaranteed to avoid bias in the test set for more generalized cases. 1 Introduction In machine learning one typically assumes that the true classification of an object depends only on the object itself and given the object, is independent of the classification of other objects. In this case, setting aside a sufficiently large and randomly chosen part of the training data as a test set, the observed sample error on the test set is an unbiased estimator of true error. A number of error measures have been proposed for the propositional domain including off-training set error OTS [Wolpert, 1996], which emphasizes error estimation on unobserved instances. However, in many application settings, those mainstream approaches to model evaluation might be inappropriate. As pointed out by [Jensen and Neville, 2002], among many others, whenever there is autocorrelation, i.e., whenever the target value of one object depends not only on the object itself, but also on other objects classifications or information that is shared between objects, observed error on a randomly chosen test set may not be an unbiased estimator anymore. Subgraph sampling as proposed by [Jensen and Neville, This paper has been presented at the 4th Multi-Relational Data Mining Workshop accompanying the 10th ACM SIGKDD 2005 copyright ACM, 2005 and as late-breaking paper at the ILP 2005. 2002] eliminates this effect but assumes that information shared in the training set does not reoccur with future instances of the domain. We introduce generalized subgraph sampling which prevents a bias due to an increased or reduced proportion of linked objects in the test set and which includes subgraph sampling as a special case. The paper is organized as follows. Section 2 gives an overview of commonly used methods for error estimation on propositional data and shows by a simple experiment the different behavior of true error and OTS error. Section 3 addresses the bias introduced by linked objects in multirelational data. We show the importance of reflecting a realistic proportion of linked objects in the test set and consequently present an adapted sampling algorithm in Section 4. Section 5 discusses related work. The paper concludes with an example of hierarchical relationships in traffic data and poses further challenges for error estimation on multirelational data. 2 Model Evaluation in Propositional Domains A major task in machine learning is function approximation from examples. Let X be an instance space with a fixed distribution D X, let Y denote a target space, let f : X Y be an unknown target function and let H be a hypothesis space consisting of a set of functions mapping X to Y. Given a training set S X Y assumed to be generated such that y = fx, the goal of function approximation is to find a hypothesis h H which best approximates f with respect to some error function error : Y Y IR and some error measure. Examples for an error function are squared error and 0-1-loss. Below we define an error measure called true error. Definition 1 True Error For a distribution D X and a target function f, the true error of a hypothesis h H is defined as error te D X h, f := E DX [errorhx, fx]. Usually, only a fraction of the instance space can be obtained as labeled data and must suffice for both model generation and model evaluation. Error estimation on the training set is widely known to result in an optimistic bias overfitting either due to noise in the training data or a coincidental relation between the target attribute and an otherwise uncorrelated attribute [Mitchell, 1997]. One possibility to eliminate this bias is to partition the data sample into two independent sets e.g of sizes 2 3, 1 3, which will be used during the training and validation period of the classifier respectively. However, a single evaluation does not account

for random variation induced by a specific training or test set as [Dietterich, 1998] points out. Therefore the evaluation process is often repeated, e.g. by random subsampling or cross validation. Random subsampling simply conducts a series of trials, providing each with a new fixed-size random split of the data sample, and averages over the estimated errors. As the training and test sets of different trials overlap, the assertion of independent experiments cannot be made. k-fold cross validation splits the data sample into k equal-sized folds, one of which is iteratively used as test set, while the union of the remaining k 1 folds serves as training data. The final error is calculated as the mean of the k conducted tests. Clearly, cross validation provides for independent test sets between the iterations. Nevertheless, any two training samples will overlap by a fraction of 1 1 k 1. An often formulated goal in inductive learning is to estimate a learner s accuracy on future instances. As the labels of training instances are known in advance, they can simply be stored in a look-up table and be reused on future encounters. Given a noise free and consistent data set, the classification of these objects will always be correct. Using true error for accuracy estimation, the influence of previously seen and stored objects increases with the size of the training set. This affects true error on basis of quantity notwithstanding to the quality a hypothesis displays for yet unseen instances. In order to circumvent this effect, [Wolpert, 1996] proposed to estimate off-training set error OTS, which excludes training instances when estimating the error of a hypothesis. Definition 2 OTS Error For a distribution D X, a target function f and a training set S, the OTS error of a hypothesis h H is defined as errord ots X h, f := E D X [errorhx, fx] { 0 if x S with P D X x = P DX x Rx / S P otherwise D X xdx defining the OTS distribution D X. For the remainder of this paper we assume that the data sample S does not contain repetitions as is likely in case of a continuous distribution with P X = x = 0 for each instance. The behavior of OTS error and true error is nearly identical when the size of the training set is very small compared to the size of the instance space [Wolpert, 1996]. The difference between both measurements increases, however, with the proportion of the training set with respect to the instance space. To illustrate this, we conducted a simple experiment on noisefree artificial data. Our instance space X = {0, 1} 10 is defined by 10 binary attributes. The binary target value Y is equally distributed P 1 = P 0 = 0.5 and assigned at random. We applied a random split and ran the experiment several times varying the split factor. A 1-nearest-neighbor classifier was used each time a computing true error on the complete instance space and b estimating OTS error with 10-fold cross validation. As cross validation partitions the data sample into a non-overlapping training and test set per iteration and each instance appears only once in the data set, cross validation effectively estimates OTS error 1. In order to eliminate random effects, we averaged results over 500 trials. 1 In the case of duplicate objects cross validation must be able to recognize those objects and to treat them as single instance in order to estimate OTS error. Mean Error over 500 Runs Proportion of Training Set 10 Fold CV True Error Figure 1: OTS error and true error on propositional data with respect to training size Figure 1 shows the difference between true error and OTS error on varying sample sizes. Both error measures start approximately at the default error of 0.5 which can be expected for unseen objects. With increasing training size the expected OTS error remains constant while the true error declines until the complete instance space is represented by the training set. [Wolpert et al., 1998] conducted studies and showed that for a Bayes-optimal learner the expected OTS error can even increase with the training set size while the expected true error is a strictly non-increasing function of the training set size. Thus, the decision which error measure to apply must be carefully considered and is application dependent. If the labels of the training instances are known with certainty and a generalization of the training instances is not required, OTS error measure is appropriate. Nevertheless, it must be provided that the entire training instances can be stored for later reuse. Contrarily, if the learner is expected to find a general regularity the estimation of true error is desired. In this case, cross validation is not an appropriate error measure and the error estimate needs to be corrected by a combination with the training set error. 3 Linkage Bias in Multirelational Domains In multirelational domains, the assumption of independent instances, which is crucial to cross validation, cannot be taken for granted. Frequently, as [Jensen and Neville, 2002] point out, the instance space is structured by a number of objects highly correlated with the target attribute. [Jensen and Neville, 2002] give a simple example involving a relational learning problem. Based on data from the Internet Movie Database 2 they consider the task of predicting whether a movie has box office receipts of more than $2 million given information about the studio that made the movie. Figure 2 shows the relevant structure of the movie data set. More formally, the application consists of two kinds of objects, movies X and studios A. We will refer to a studio a A as a neighbor of a movie x X if the studio is the producer of that movie. The set of movies sharing the same studio a forms the neighborhood NB a of a. The degree δ a of a is specified as the cardinality of its neighborhood NB a. All movies within a neighborhood are related to each other. To clearly identify the effect of autocorrelation, [Jensen and Neville, 2002] described each studio by a set of ten at- 2 http://www.imdb.com/

Figure 2: Internal structure of the movie data tributes, the values of which were generated randomly for each studio. In their experiments they performed random split of movies into training and test data, each time including the movie together with information about the producing studio in the respective set. Clearly, a movie appears either in the training or the test set. However, since there are many movies and only a few studios, for most random splits movies from almost all studios will appear in both training and test set. Given that the random attributes on studios are the only information given to the learning system and are generated independent of the classification of their associated movies it seems very plausible that, as argued in [Jensen and Neville, 2002], the true error of any model should equal the baseline error of a classifier that simply always predicts the majority class this error being 0.45 in the particular data set of [Jensen and Neville, 2002]. However, when using random test sets to measure the error of a simple relational learner FOIL, the observed error was significantly lower than this baseline value. When the experiment was repeated with the same data, however randomly assigning the class labels to movies, observed sample error closely corresponded to the baseline error of a majority class classifier. Conducting a series of experiments using random sampling, [Jensen and Neville, 2002] showed that the bias increases with the so-called linkage, representing the number of movies made by a single studio, and autocorrelation in the data set. Let us think about the reasons for the lower observed error compared to the default error. Since there are only very few studios in the data set, with randomly generated values of 10 attributes, the values of these 10 attributes effectively form a key that uniquely identifies each studio. In fact, with so few studios, each studio can be uniquely identified with simple selector conditions on one or two attributes. Thus, even though no studio identifiers are given, in the experiment of [Jensen and Neville, 2002] FOIL can build rules of the type if the studio is studio-n, then class is +. Knowing that big studios will make many movies with big box office receipts, it is very plausible that such rules perform better than guessing 3. Under which circumstances should we consider this behavior as overfitting? Clearly, such rules will be of no use when classifying a movie made by a studio that previously has never appeared in our data. This has led [Jensen and Neville, 2002] to postulate that the dependence between the training and test set should be eliminated. They present a procedure, subgraph sampling, which ensures that any information in this case studios shared between different instances is included in either the training or the test set, 3 Nevertheless, it may be argued that even though semantically useful the syntactic appearance of those rules is less intuitive than the previously stated example. A user may easily mistake them for a general regularity and one should aim at a clear representation if a rule indeed forms a unique studio identifier. but not in both, thus eliminating the above mentioned bias in error estimation. Looking more closely, however, shows that this approach considers only the case in which future movie objects possess a yet unknown studio as neighbor. In many applications, and in this one in particular, the average error made by a given hypothesis on future instances, yet, depends on both instances related to objects in the training set as well as on instances with still unknown neighbors. In the movie data set it is actually far more likely that a new movie will be produced by one of the already existing studios than by a newly founded studio. If error estimation is solely based on the latter kind of movies the error will be overestimated. Instead, the test set should reflect the probability of future encounters with instances related to objects in the training set. We will refer to this probability as the known neighbor probability. Definition 3 Known Neighbor Probability For an instance space X, a distribution D X, a sample S, a set A of neighbors and a function nb : X A assigning to each instance x X its neighbor a A, the known neighbor probability is defined as p kn S := P nbx s S nbs with x D X. In conformance with the OTS approach, p kn S denotes the probability that an instance x drawn according to the OTS distribution D X shares information with objects of the data sample. Given a sample S, the known neighbor probability p kn S is a domain property. If the distribution over the complete instance space were known, the computation of p kn S would be straightforward. As this is almost never the case usually all we have is the sample S, the known neighbor probability must be supplied by the user based on application considerations. The training set S does not allow to infer the known neighbor probability directly, because the neighbors for all given instances are known. In particular, simply performing a training/test split does not produce the right known neighbor frequency in the test set. Actually, the frequency of known neighbors in the test set varies depending on the training/test split using random sampling. This behavior defeats any attempt for a proper estimation of p kn S from the training set. In order to illustrate the effect and its influence on error estimation, let us conduct another small experiment. Continuing the example of [Jensen and Neville, 2002] we created an artificial data sample X with 1000 movies and 20 studios A, each studio a A producing 50 movies. The labels of the movies were drawn from a binary uniform distribution, and the studios were identified by a number of 10 randomly chosen binary attributes. We introduced autocorrelation such that 80 per cent of the movies for a given studio possessed labels of one class. Applying random sampling we varied the training/test split and calculated the frequency of known neighbors in the test set. We attached the attributes of the neighboring studio to each movie and used a Weka J48 tree [Witten and Frank, 2000] with default parameters for classification. Figure 3 shows the averaged results over 1000 trials. Clearly, with an increasing proportion of training examples the number of studios in the training set increases as well. As the decision tree is able to find rules to distinguish studios by their attributes, and the class labels of

Mean Error over 1000 Runs Training/Test Split Proportion of Training Set Figure 3: Classification results on artificial movie data using random sampling a frequency of known neighbors in the test set b error and standard deviation measured on the test set movies from the same studio are highly correlated, the estimated error decreases with each additionally known studio. The measured error thus depends on the chosen split factor. The experiment shows another interesting aspect with regard to the variance of the error. When the average number of known neighbors reaches its saturation at a training/test split of approximately 0.16, the variance of the error decreases still further. This can be explained by two effects, first, the maximum number of represented studios stabilizes with further training examples. Second, the weight of random effects due to incomplete autocorrelation decreases with the number of available movies per studio. In the next section we propose generalized subgraph sampling which ensures unbiased error estimation independent of the training/test split. 4 Generalized Subgraph Sampling Sampling procedures for relational data pursue two major approaches in recent work. The first strategy exploits natural divisions to create independent training and test sets. For an example, the WebKB 4 data set provided by IPL 98 contains web pages collected from computer science departments of four universities. As the departments do not link to each other, the data can be split to include web pages from three universities in the training set and web pages of the fourth university in the test set. The second strategy is based on temporal aspects of the data. Data collected over time may be subdivided into sets of non-overlapping consecutive periods of time. Each of these sets can be used 4 http://www-2.cs.cmu.edu/afs/cs.cmu.edu/project/theo- 20/www/data/ as training input while testing is performed on the set of the following period. [Neville and Jensen, 2003] apply this technique to data from the Internet Movie Database separating movies by their year of appearance. Thus, each test set provides a natural proportion of known and unknown studios even though the portion of each studio might vary from year to year. If neither a natural nor a temporal split can be applied to the data, the sampling procedure must be adjusted to introduce known neighbor probability as required. Generalized subgraph sampling is a sampling procedure which is based on the known neighbor probability and which includes subgraph sampling as proposed by [Jensen and Neville, 2002] for the special case of p kn S = 0. Generalized subgraph sampling ensures that for a given data sample S, a percentage p train for the training/test split, and the known neighbor probability p kn S with respect to objects A, the resulting test set S test reflects the same proportion of known neighbors with regard to the training set S train as specified by the known neighbor probability. In order to simplify notation, we use the following conventions: A subset is indicated by adding a subscript to the name of the originating set, e.g. S test denotes the test set created from sample S, and S a denotes the subset of all instances in S that are linked to object a, i.e. the neighborhood of a in S. p train = 1 p test denotes the proportion of training examples of a given training/test split. p test,un = p test 1 p kn is the desired proportion of objects with unknown neighbors in the test set relative to the sample size. n, n test, n test,un denote the size of the sample, the test set and the number of objects with unknown neighbors in the test set respectively. Algorithm 1 calculates first based on the size of the sample and the given percentages p train and p kn S the required number of objects with unknown neighbors in the test set n test,un. The disposable space is filled by successively choosing, among the remaining shared objects, a neighbor with the smallest degree and placing its neighborhood in S test. As it cannot be guaranteed that the number of obtained objects reaches exactly n test,un, the size of the test set is increased in order to maintain the required known neighbor probability thus violating the proportion of training/test split. The remaining part of the test set is populated ensuring that for each chosen instance a neighbor is placed in the training set. Algorithm 1 has two obvious drawbacks. As the number of required objects with unknown neighbors in the test set is approximated by selecting from a size-based ranking, first, the algorithm prefers instances belonging to neighbors with a small degree. Second, the selection may reach shared objects with a high degree and is thus likely to exceed the bound n test,un of unknown neighbors by a large number. The adjustment of the training/test split then may lead to very small training sizes. We therefore propose an adaptation of the selection procedure which substitutes lines 5-11 of algorithm 1. Algorithm 2 pursues a random sampling strategy in order to choose the next neighbor which is unknown to the training set. In contrast to algorithm 1 excessive objects with unknown neighbors are discarded such that the final partition of S still conforms to p train and p kn S. Instead, the

Algorithm 1 GENERALIZED SUBGRAPH SAMPLING tbh[0] Input: sample S, percentage p train for the train/test split, known neighbor probability p kn S with respect to objects A Output: training set S train, test set S test 1: S train = S test = 2: n test = S 1 p train 3: n test,un = S 1 p train 1 p kn 4: compute the set A of neighbors and the neighborhood S a for each a A 5: while S test < n test,un do 6: choose a A with the smallest degree δ a 7: S test = S test S a 8: S = S\S a 9: A = A\{a} 10: end while 11: n test = S test /1 p kn 12: while S test < n test do 13: choose a A randomly 14: if S a 2 then 15: choose two objects s a1, s a2 from S a randomly 16: S test = S test {s a1 } 17: S train = S train {s a2 } 18: S = S\{s a1, s a2 } 19: else 20: S train = S train S a 21: S = S\S a 22: A = A\{a} 23: end if 24: end while 25: S train = S train S 26: return S train, S test Algorithm 2 RANDOM SELECTION OF UNKNOWN NEIGHBOR TEST SET tbh[0] 1: n = S 2: while S test < n test,un do 3: choose a A randomly 4: S test = S test S a 5: S = S\S a 6: A = A\{a} 7: end while 8: r excess = S test n test,un /n 9: n = n 1 r excess /p train + p kn S p trainp kn 10: n test = n 1 p train 11: n test,un = n 1 p train 1 p kn 12: while S test > n test,un do 13: choose s t S test randomly 14: S test = S test \{s t } 15: end while sizes of the training and test set are decreased proportionally. This leads to a recursive adaptation of the number of available samples because the proportion r excess of excessive objects is calculated based on the sample size n. The exclusion of objects with unknown neighbors reduces the total number of available objects and consequently reduces n test,un, leading again to the exclusion of objects with unknown neighbors. The final number n of samples used can be calculated as inf n = n 1 r excess p test,un n=1 = n 1 r excess 1 p test,un = n 1 p train + p kn S r excess p train p kn S In the above equation r excess = S test n test,un /n denotes the fraction of excessively selected objects with unknown neighbors and p test,un = 1 p train 1 p kn serves as substitution for the desired proportion of unknown neighbors, which has been reversed in line three. Based on the new sample size n, the final sizes of the training and test set are determined. The adapted sampling procedure in algorithm 2 maintains the given split proportion and known neighbor probability. Yet, information is lost as several instances of the data sample are discarded. 5 Related Work Recent work has examined numerous techniques which can exploit the relationship between neighboring objects in relational data. Among those are Probabilistic Relational Models PRMs [Getoor et al., 2001], Relational Markov Networks [Taskar et al., 2002] and Relational Density Networks [Neville and Jensen, 2003], which have developed from Bayesian Networks, Markov Networks and Dependency Networks respectively. Each model captures special types of dependencies inherent to the data. If a model is able to represent dependencies between class labels of related objects, algorithms for collective inference can be applied [Jensen et al., 2004]. More specifically, collective classification is the simultaneous decision on the class labels of all instances exploiting correlation between the labels of related instances [Jensen et al., 2004; Taskar et al., 2002]. For example, a simple Relational Neighbor RN classifier, which estimates class probabilities solely based on labels of related objects, has been proposed by [Macskassy and Provost, 2003]. Yet, collective classification involves not only predictions based on known labels, but also includes strategies to exploit labels that can be inferred. Several approaches, including [Macskassy and Provost, 2003; Slattery and Mitchell, 2000], conduct information propagation by iteratively labeling instances. Studies on classifiers applying collective inference have shown that those techniques significantly reduce classification error [Macskassy and Provost, 2003; Taskar et al., 2001]. Each of these models has a particular way of dealing with a possible bias introduced by linked objects. It is a topic of future work to investigate the precise relationship between those approaches and generalized subgraph sampling as advocated in this paper..

6 Conclusion and Future Work In multirelational domains it is well known that linkage introduced by shared objects and autocorrelation causes a bias in test procedures. We demonstrated that the bias varies with the training/test split of a given sample and showed that the test set should reflect the known neighbor probability as encountered on future instances. We presented a sampling procedure, generalized subgraph sampling, which guarantees to partition a sample such that the test set reflects a given known neighbor probability and thus avoids a bias in the error estimate. While this work concentrated on problem identification of hypothesis evaluation in relational domains and proposed and a first sampling procedure, future studies will include an experimental or theoretical evaluation of the proposed technique. Other interesting research topics identified during our work concern the question whether it is possible to estimate the known neighbor probability from the data sample under certain conditions. Further, the exploration of corrected sampling procedures applicable to cross validation seem very valuable to us. Figure 4: Hierarchical structure of traffic data This paper is motivated by work on traffic data. Traffic data provides information about the number of vehicles and pedestrians that pass along a street during a fixed time interval. The instance space hereby consists of street segments, which denote the part of a street between two intersections. Each segment can be characterized by numerous attributes, e.g. the street name and type, its spatial coordinates and nearby points of interest. As the traffic frequencies within one street are very similar, instances belonging to the same street are subject to autocorrelation. Likewise the expected number of vehicles traveling on a street in a capital city like Berlin is much higher than in some small village. Thus, a hierarchical structure of linkage is imposed on the domain, which is depicted in Figure 4. In future research we will therefore investigate the behavior of linkage and autocorrelation in hierarchical and graph structured data and try to deduce corrected sampling procedures for those domains. classification. In Proc. of the 10th ACM SIGKDD International Confererence on Knowledge Discovery and Data Mining. [Macskassy and Provost, 2003] Macskassy, S. and Provost, F. 2003. A simple relational classifier. In Proc. of the 2nd Multi-Relational Data Mining Workshop, 9th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. [Mitchell, 1997] Mitchell, T. M. 1997. Machine Learning, pages 66 72. McGraw-Hill, New York. [Neville and Jensen, 2003] Neville, J. and Jensen, D. 2003. Collective classification with relational dependency networks. In Proc. of the 2nd Multi-Relational Data Mining Workshop, 9th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. [Slattery and Mitchell, 2000] Slattery, S. and Mitchell, T. 2000. Discovering test set regularities in relational domains. In Proc. of the 17th International Conference on Machine Learning. [Taskar et al., 2002] Taskar, B., Abbeel, P., and Koller, D. 2002. Discriminative probabilistic models for relational data. In Proc. of the 18th Conference on Uncertainty in Artificial Intelligence. [Taskar et al., 2001] Taskar, B., Segal, E., and Koller, D. 2001. Probabilistic classification and clustering in relational data. In Proc. of the 17th International Joint Conference on Artificial Intelligence. [Witten and Frank, 2000] Witten, I. H. and Frank, E. 2000. Data Mining - Practical Machine Learning Tools and Techniques with Java Implementations. Morgan Kaufmann, San Francisco. [Wolpert, 1996] Wolpert, D. H. 1996. The lack of a priori distinctions between learning algorithms. Neural Computation, 87:1341 1390. [Wolpert et al., 1998] Wolpert, D. H., Knill, E., and Grossmann, T. 1998. Some results concerning off-trainingset and iid error for the gibbs and bayes optimal generalizers. Statistics and Computing, 81:35 54. References [Dietterich, 1998] Dietterich, T. G. 1998. Approximate statistical test for comparing supervised classification learning algorithms. Neural Computation, 107:1895 1923. [Getoor et al., 2001] Getoor, L., Friedman, N., Koller, D., and Pfeffer, A. 2001. Relational data mining. In Dzeroski, S. and Lavrac, N., editors, Learning Probabilistic Relational Models, pages 307 335. Springer-Verlag, Berlin. [Jensen and Neville, 2002] Jensen, D. and Neville, J. 2002. Autocorrelation and linkage cause bias in evaluation of relational learners. In Proc. of the 12th International Conference on Inductive Learning. Springer- Verlag. [Jensen et al., 2004] Jensen, D., Neville, J., and Gallagher, B. 2004. Why collective inference improves relational