A Modular Reduction Method for k-nn Algorithm with Self-recombination Learning

Similar documents
A Modular k-nearest Neighbor Classification Method for Massively Parallel Text Categorization

Efficient Pairwise Classification

Efficient Pruning Method for Ensemble Self-Generating Neural Networks

Minimal Test Cost Feature Selection with Positive Region Constraint

A Support Vector Machine Classifier with Automatic Confidence and Its Application to Gender Classification

Fast k Most Similar Neighbor Classifier for Mixed Data Based on a Tree Structure

Combination of PCA with SMOTE Resampling to Boost the Prediction Rate in Lung Cancer Dataset

K-Means Clustering With Initial Centroids Based On Difference Operator

Color-Based Classification of Natural Rock Images Using Classifier Combinations

Dynamic Clustering of Data with Modified K-Means Algorithm

Using the Geometrical Distribution of Prototypes for Training Set Condensing

Multiple Classifier Fusion using k-nearest Localized Templates

KBSVM: KMeans-based SVM for Business Intelligence

A Genetic k-modes Algorithm for Clustering Categorical Data

Gender Classification Based on Support Vector Machine with Automatic Confidence

Using Decision Boundary to Analyze Classifiers

Using a genetic algorithm for editing k-nearest neighbor classifiers

WEIGHTED K NEAREST NEIGHBOR CLASSIFICATION ON FEATURE PROJECTIONS 1

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

SOMSN: An Effective Self Organizing Map for Clustering of Social Networks

Available online at ScienceDirect. Procedia Computer Science 35 (2014 )

Leave-One-Out Support Vector Machines

Naïve Bayes for text classification

Nearest Cluster Classifier

ORT EP R RCH A ESE R P A IDI! " #$$% &' (# $!"

APPLICATION OF THE FUZZY MIN-MAX NEURAL NETWORK CLASSIFIER TO PROBLEMS WITH CONTINUOUS AND DISCRETE ATTRIBUTES

Figure (5) Kohonen Self-Organized Map

Efficient Pairwise Classification

An Empirical Study of Hoeffding Racing for Model Selection in k-nearest Neighbor Classification

Feature-weighted k-nearest Neighbor Classifier

PARALLEL SELECTIVE SAMPLING USING RELEVANCE VECTOR MACHINE FOR IMBALANCE DATA M. Athitya Kumaraguru 1, Viji Vinod 2, N.

I. INTRODUCTION II. RELATED WORK.

Support Vector Regression for Software Reliability Growth Modeling and Prediction

BRACE: A Paradigm For the Discretization of Continuously Valued Data

Global Metric Learning by Gradient Descent

Flexible-Hybrid Sequential Floating Search in Statistical Feature Selection

Fuzzy Ant Clustering by Centroid Positioning

ICA as a preprocessing technique for classification

FUZZY KERNEL K-MEDOIDS ALGORITHM FOR MULTICLASS MULTIDIMENSIONAL DATA CLASSIFICATION

STUDYING OF CLASSIFYING CHINESE SMS MESSAGES

Some questions of consensus building using co-association

Using Association Rules for Better Treatment of Missing Values

The Role of Biomedical Dataset in Classification

CloNI: clustering of JN -interval discretization

EFFICIENT ATTRIBUTE REDUCTION ALGORITHM

Why Do Nearest-Neighbour Algorithms Do So Well?

Class-Information-Incorporated Principal Component Analysis

A Novel Template Reduction Approach for the K-Nearest Neighbor Method

Determination of Similarity Threshold in Clustering Problems for Large Data Sets

The Potential of Prototype Styles of Generalization. D. Randall Wilson Tony R. Martinez

Nearest Cluster Classifier

A Fuzzy C-means Clustering Algorithm Based on Pseudo-nearest-neighbor Intervals for Incomplete Data

Data Mining and Data Warehousing Classification-Lazy Learners

Enhancing K-means Clustering Algorithm with Improved Initial Center

Two-step Modified SOM for Parallel Calculation

INF4820 Algorithms for AI and NLP. Evaluating Classifiers Clustering

Parallel Evaluation of Hopfield Neural Networks

Resource Load Balancing Based on Multi-agent in ServiceBSP Model*

Image Classification Using Wavelet Coefficients in Low-pass Bands

Learning the Three Factors of a Non-overlapping Multi-camera Network Topology

An Improved KNN Classification Algorithm based on Sampling

Relation Organization of SOM Initial Map by Improved Node Exchange

Content Based Image Retrieval system with a combination of Rough Set and Support Vector Machine

Fuzzy Bidirectional Weighted Sum for Face Recognition

Reihe Informatik 10/2001. Efficient Feature Subset Selection for Support Vector Machines. Matthias Heiler, Daniel Cremers, Christoph Schnörr

Use of Multi-category Proximal SVM for Data Set Reduction

Using K-NN SVMs for Performance Improvement and Comparison to K-Highest Lagrange Multipliers Selection

CLUSTERING BIG DATA USING NORMALIZATION BASED k-means ALGORITHM

Image Segmentation Based on Watershed and Edge Detection Techniques

K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors

A fuzzy k-modes algorithm for clustering categorical data. Citation IEEE Transactions on Fuzzy Systems, 1999, v. 7 n. 4, p.

HALF&HALF BAGGING AND HARD BOUNDARY POINTS. Leo Breiman Statistics Department University of California Berkeley, CA

Time Series Prediction as a Problem of Missing Values: Application to ESTSP2007 and NN3 Competition Benchmarks

An Effective Performance of Feature Selection with Classification of Data Mining Using SVM Algorithm

organizing maps generating abilities of as possible the accuracy of for practical classifier propagation and C4.5) [7]. systems.

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation

Introducing Partial Matching Approach in Association Rules for Better Treatment of Missing Values

Cluster based boosting for high dimensional data

Kernel Combination Versus Classifier Combination

Non-linear gating network for the large scale classification model CombNET-II

Estimating Feature Discriminant Power in Decision Tree Classifiers*

A Comparison of Global and Local Probabilistic Approximations in Mining Data with Many Missing Attribute Values

The Effects of Outliers on Support Vector Machines

Estimating Missing Attribute Values Using Dynamically-Ordered Attribute Trees

Using Decision Trees and Soft Labeling to Filter Mislabeled Data. Abstract

Locality Preserving Projections (LPP) Abstract

C-NBC: Neighborhood-Based Clustering with Constraints

Semi-Supervised Clustering with Partial Background Information

A Graph Theoretic Approach to Image Database Retrieval

AN IMPROVED HYBRIDIZED K- MEANS CLUSTERING ALGORITHM (IHKMCA) FOR HIGHDIMENSIONAL DATASET & IT S PERFORMANCE ANALYSIS

Efficient Tuning of SVM Hyperparameters Using Radius/Margin Bound and Iterative Algorithms

Improving the Efficiency of Fast Using Semantic Similarity Algorithm


AN IMPROVEMENT TO K-NEAREST NEIGHBOR CLASSIFIER

SoftDoubleMinOver: A Simple Procedure for Maximum Margin Classification

Self-Organizing Maps for Analysis of Expandable Polystyrene Batch Process

Training Data Selection for Support Vector Machines

CIS UDEL Working Notes on ImageCLEF 2015: Compound figure detection task

A Boosting-Based Framework for Self-Similar and Non-linear Internet Traffic Prediction

A New Type of ART2 Architecture and Application to Color Image Segmentation

Transcription:

A Modular Reduction Method for k-nn Algorithm with Self-recombination Learning Hai Zhao and Bao-Liang Lu Department of Computer Science and Engineering, Shanghai Jiao Tong University, 800 Dong Chuan Rd., Shanghai 200240, China bllu@sjtu.edu.cn Abstract. A difficulty faced by existing reduction techniques for k-nn algorithm is to require loading the whole training data set. Therefore, these approaches often become inefficient when they are used for solving large-scale problems. To overcome this deficiency, we propose a new method for reducing samples for k-nn algorithm. The basic idea behind the proposed method is a self-recombination learning strategy, which is originally designed for combining classifiers to speed up response time by reducing the number of base classifiers to be checked and improve the generalization performance by rearranging the order of training samples. Experimental results on several benchmark problems indicate that the proposed method is valid and efficient. 1 Introduction Given a training set of previously labeled samples and an unknown sample x, the k-nearest neighbor (k-nn) algorithm assigns the most frequently represented class-label among the k closest prototypes to x. Thek-NN algorithm and its derivatives have been shown to perform well for pattern classification in many domains over the last 40 years. In theory, the k-nn algorithm is asymptotically optimal in the Bayesian sense [1]. In other words, k-nn performs as well as any other classifier, provided that there is an arbitrarily large number of (representative) prototypes available and the volume of the k-neighborhood of x is arbitrarily close to zero for all x. Basically, the k-nn algorithm needs the entire training set as the representative set. Therefore, the space requirement of storing the complete set and the high computational cost for the evaluation of new samples are often criticized. The k-nn algorithm also shows sensitivity to outliers, i.e. noisy or even erroneously labeled prototypes. To overcome these drawbacks, various techniques To whom correspondence should be addressed. This work was supported by the National Natural Science Foundation of China under the grants NSFC 60375022 and NSFC 60473040. J. Wang et al. (Eds.): ISNN 2006, LNCS 3971, pp. 537 544, 2006. c Springer-Verlag Berlin Heidelberg 2006

538 H. Zhao and B.-L. Lu have been developed. Two main types of algorithms can be identified: prototype generation and prototype selection. The first group focuses on merging the initial prototypes into a small set of prototypes, so that the performance of the k-nn algorithm is optimized. Examples of such techniques are k-means algorithm [2] and learning vector quantization algorithm [3]. The second group aims at the reduction of the initial training set and/or increasing the accuracy of the NN prediction. Various editing and condensing methods were developed, such as condensing algorithm [4] and editing algorithm [5]. Wilson et al. recently gave a good review for all these reduction techniques for the k-nn algorithm [6]. All of these existing reduction algorithms, however, still need to store the whole training set firstly to meet the requirement of fast processing. The reason is that all of them are actual k-nn based self-test of the training set. Therefore, if there is any no parallel or modular techniques introduced to the k-nn algorithm, those existing k-nn reduction techniques can not efficiently perform reduction of a large-scale data set. A modular k-nn algorithm has been proposed in our previous work [7]. Of course, this modular k-nn algorithm is mainly suitable to dealing with very large-scale data sets, and a user can not divide a training data set into too many subsets in order to maintain a stable combining accuracy. Otherwise, some assistant techniques, such as clustering, should be introduced to help the partition of the training data set [8][9]. However, all clustering procedures also require additional computing time and the same storage space unless there is a parallel or modular version of clustering algorithm. In one word, unless we can directly solve the reduction problem for the k-nn algorithm or there exists a method which can carry out both partition and reduction at the same time, we will face many other concomitant problems. In this paper, s self-recombination learning scheme based reduction method is proposed for overcoming above difficulties. The proposed method is based on combining classifier from the quadratic decomposition of two-class task. It is also a framework of parallel or modular self-test based on the k-nn algorithm. 2 Combining Classifier Since any multi-class classification problem can be solved by using two-class classification techniques according to well known one-against-rest and oneagainst-one task decomposition methods. Especially, the one-against-one task decomposition method has been an innate parallel and modular processing method for binary classifier modules. Thus, we will only concern with two-class combining classifier in the remained part of this paper. We firstly discuss the decomposition of a two-class classification task. Let X + and X be the given positive and negative training data sets, respectively, for a two-class classification task T, X + = {(x + i, +1}l+ i=1, X = {(x i, 1)}l i=1 (1)

A Modular Reduction Method for k-nn Algorithm 539 where x + i R n and x i R n are the input vectors, and l + and l denote the numbers of positive training samples and the number of negative training samples, respectively. X + and X can be partitioned into N + and N subsets respectively, X + j = {(x +j i, +1)} l+ j i=1, j =1,..., N+ (2) X j = {(x j i, 1)} l j i=1, j =1,..., N where N + j=1 X + j = X +,1 N + l +,and N j=1 X j = X,1 N l. Matching all these subsets, X + i and X j, where i = 1,..., N+ and j = 1,..., N,canyieldN + N relatively smaller and more balanced (if needed) two-class subsets pairs. Each of them can be expressed as T (i,j) = X + i X j,i=1,..., N+,j =1,..., N (3) Assign a k-nn classifier as a base classifier to learn from each subset T (i,j). In the learning phase, all of the k-nn base classifiers are independent from each other and can efficiently work in a parallel way. We also denote the corresponding k-nn base classifier as T (i,j), if there is not any misunderstanding. For the convenient, these base classifiers, T (i,j),j =1,,N, are defined as positive group and i is defined as its group label, and those base classifiers, T (i,j),i=1,,n +, are defined as negative group and j is defined as its group label. Intuitively, if all classifiers in a positive group support the decision output of positive class for an unknown input sample (notice that this is a very strong condition), then to assign this unknown sample the positive class label is very natural. It is the same for a negative sample and its related negative group. We call a positive group whose all member base classifiers support the positive class as a positive winning group and a negative group whose all member base classifiers support the negative class as a negative winning group. It should be noted that it is impossible that positive and negative winning groups exist at the same time, because these two groups must share one same base classifier and this classifier can not output two different classification results at the same time. To finish a combination without any preference, we define the combination strategy as follows. The combining output of the positive class is determined by the condition that no negative winning groups exist, and the combining output of the negative class is determined by the condition that no positive winning groups exist. Sometimes, neither positive nor negative winning group exits in application, that is the reason why we make the definitions with the negatory conditions. In order to effectively handle the case that no winning group exists and realize an efficient combination, the following symmetrical selection algorithm is given as a practical realization of the above combining strategy.

540 H. Zhao and B.-L. Lu 1) Set the initial labels, i =1andj =1. 2) Repeat the following operations: (a) Check the base classifier T (ij). (b) If T (ij) supports the positive class, then j = j +1. (c) If T (ij) supports the negative class, then i = i +1. (d) If i = 1 + N +, then the combining output is negative class and the algorithm ends here. (e) If j = 1 + N, then the combining output is positive class and the algorithm ends here. We now show that the number of checked base classifiers in symmetrical selection algorithm for each input sample will be never larger than N + +N 1. As an intuitive expression, we imagine that all outputs of base classifiers are represented in a binary matrix with N columns and N + rows. It is obvious that the algorithm is equally a search procedure in the matrix, in which the start point is the top-left corner and the end point is near the bottom-right corner. Each checking for a base classifier means one row or one column in the matrix must be excluded. The search direction will turn to the next row or column without any backtracking after each checking. Consider the maximumlength search from the top-left corner to the bottom-right corner in the matrix. This maximum-length search will cover N + + N 1 elements. That is, the number of checked base classifiers in symmetrical selection for an input sample will not be larger than N + + N 1, which is much less than the number of all base classifiers, N + N. The theoretical analysis in our existing study has actually given a probability model for symmetrical selection based combination through the performance estimation of two-class combining classifier [10]. 3 Reduction with Self-recombination Learning Self-recombination learning (SRL) originates such a simple idea. Consider the searching procedure of symmetrical selection algorithm in the outputs matrix of all base classifiers, the start point is the base classifier in the top-left corner, and the end point will be located in the most right column or the most below rows. We may notice if a base classifier is nearer to the bottom-right point of the output matrix, it will have the less chance to be checked and the less capability to determine the final combination output. Thus, we may obtain the higher combining accuracy by arranging those base classifiers with higher error rate to the more right columns and the more below rows. Finally, we may remove the learning of those base classifiers from bad samples to realize the sample reduction. The proposed reduction algorithm with self-recombination learning can be described as follows. 1) Specify the volumes of all subsets. 2) Set the maximum turns of self-learning, M, and the counter of the selflearning times, t =1.

A Modular Reduction Method for k-nn Algorithm 541 3) While t<m,do (a) Extract specified number of training samples in order from either training set to yield all training subsets of individual class. (b) Record the classification outputs of all k-nn base classifiers for all samples in the training set. (c) Record the combining outputs of all samples of the training set according to the symmetrical selection combination. (d) Put all incorrectly classified samples in the rear part of the set for both positive and negative training sets. 4) Remove all k-nn base classifiers learning from incorrect classified samples. There are still two potential problems remained in the above self-learning scheme. One is how to determine the stop criterion of SRL. We experimentally suggest that a stable training accuracy can be taken as such a criterion. The other is how to rearrange the correctly classified samples in proper order. Always, we can finish these two tasks in one step, moving the correctly or incorrectly classified samples to the corresponding parts and rearranging the order of the samples in either part. We call such procedure as moving and rearranging algorithm (MRA). Assume the number of all samples is N Max, a suitable MRA for the less exchange times and the relatively stable learning can be described as follows. 1) Set the serial number of the sample to be handled, i =1. 2) While i<= N Max,do (a) If the sample i is correctly classified, then i = i +1.Otherwise,do i. Set the counter, C = 0, and the serial number, j = i +1. ii. If the sample j is correctly classified, then exchange samples i and j and goto a), Otherwise, let j = j +1andC = C +1. iii. If C = N Max j, then the algorithm ends here. Otherwise, go to ii.. 4 Experiments Two data sets as shown in Table 1 from UCI Repository [11] and STATLOG benchmark repository [12] are chosen for this study. To carry out decomposition of the training data set, the smaller class is divided into some equal size subsets, and the larger class is also equally divided into some subsets with the same size as the subsets of smaller classes. In k-nn algorithm, the optional values of k are set to 7, 13, 19, and 25, respectively. Comparisons of classification accuracy after reduction through multi-turn selfrecombination learning on every data sets are shown in Figs. 1. The experimental results show that the more stable and higher combining accuracy appears as the iteration of self-recombination learning increases. To show the classification accuracy of reduction is superior to the classification accuracy of using the whole samples, we give a comparison study on two cases. The experimental results shown in Fig. 2 verifies the fact that the reduction is effective on Census income data set. Actually, the comparison results are the

542 H. Zhao and B.-L. Lu Table 1. Distributions of the samples in each class on all data sets Data Set Number of Training Samples Number of Test Samples #Input Total Positive Negative Total Positive Negative Breast cancer 20000 14118 5882 7700 5482 2218 9 Census income 20107 15102 5005 10055 7552 2503 14 82 95 80 Classification Acccuracy(%) 78 76 74 72 70 68 Initial Iter=1 Iter=2 Iter=3 Iter=11 Iter=12 Iter=23 Iter=24 Classification Acccuracy(%) 90 85 80 75 70 Initial Iter=1 Iter=2 Iter=3 Iter=4 2 3 4 5 6 7 8 9 10 11 Number of Modules in Smaller Class (a) 2 3 4 5 6 7 8 9 10 11 Number of Modules in Smaller Class (b) Fig. 1. Classification accuracy after reduction through multi-turn self-recombination learning and different scale partition. (a) Census income data set, where k =7and(b) Breast Cancer data set, where k = 25. Classification Acccuracy(%) 82 80 78 76 74 Reduce Iter=1 Reduce Iter=2 72 Reduce Iter=3 Reduce Iter=4 70 Remain Iter=1 Remain Iter=2 Remain Iter=3 68 Remain Iter=4 2 3 4 5 6 7 8 9 10 Number of Modules in Smaller Class Classification Acccuracy(%) 84 82 80 78 76 74 72 70 k=7: Test Accuracy k=7: Training Accuracy k=13: Test Accuracy k=13: Training Accuracy k=19: Test Accuracy k=19: Training Accuracy k=25: Test Accuracy k=25: Training Accuracy 68 1 3 5 7 9 11 13 15 17 19 21 23 Iterative Times of Self Recombination Learning Fig. 2. Comparison of test accuracy with reduction and without reduction on Census income data set, k =7 Fig. 3. Comparison of training accuracy (storage percentage) and test accuracy in Census income data set, the smaller class is divided into 8 parts same for the other values of k on Census income and other data sets. However, the comparison results for the other two data sets are very close, since their original classification accuracies have been very high. Comparisons of test accuracy with reduction and training accuracy (also is the storage percentage after reduction) after multi-turn self recombination learning on every data sets are shown in Fig. 3. The experimental results show the con-

A Modular Reduction Method for k-nn Algorithm 543 sistent trend for the training accuracy and the test accuracy, which suggest that a stable training accuracy does be taken as the stop criterion of self learning. 5 Related Work We use a quadratic decomposition of two-class training data set, that is, the number of produced subsets or base classifiers are the product of each number of single-class subsets. Min-max modular (M 3 ) classifier [13] is also a kind of quadratic decomposition based combining classifier. The difference between the proposed combining classifier and M 3 classifier is in the combination strategy. Min-max combination in M 3 classifier needs to check O(N + N )baseclassifiers for an input sample, while symmetrical selection combination only needs N + + N 1 at most. In addition, the min-max combination prefers the output of negative class label for the reason that the min-max combining procedure is asymmetrical for two different classes. As for the combining accuracy, our experimental results show two combination strategies are very similar at most cases, and symmetrical selection combination is some better at some unbalanced cases. Surely, the quadratic decomposition is not a common task decomposition method. Most combining classifiers are based on a linear decomposition of training data set, that is, the whole training data set is divided into some parts [14]. This kind of decomposition produces less subsets and base classifiers. However, it will lose some flexibility to control ratio of the numbers of samples among every classes, while the quadratic decomposition is just on the contrary. Of course, linear decomposition can not provide the mechanism of the reduction method developed in this paper. Some recent works are also concerned with modular reduction [15]. The difference between our work and the existing approach is that it only considers an extreme case that each subset contains only two different samples and this approach is only suitable for 1-NN reduction, while our scheme is more general. In the extreme case, the combining classifier in [15] equally performs like a 1- NN classifier. Therefore, it is not strange that self-recombination learning based reduction in such case is equal to edited nearest neighbor rule of Wilson [16]. 6 Conclusions A modular and parallel reduction method with self-recombination learning for k-nn algorithm has been proposed in this paper. One of the most important features of the proposed method is that the adjustment of partitioned training data set and the reduction task can be done at the same time, while the scales of all decomposed training subsets can always be kept unchanged. Thus, the proposed method is also an effective modular k-nn algorithm. As for the future work, we expect for a more quick and stable moving and rearrangement algorithm to improve self-recombination learning. More effective stop criterion of SRL is required, too.

544 H. Zhao and B.-L. Lu References 1. Dasarathy, B. V. (ed.): Nearest Neighbor (NN) Norms: NN Pattern Classification Techniques. Los Alamitos, CA: IEEE Computer Society Press. (1991) 2. MacQueen, J. B.: Some Methods for Classification and Analysis of Multi Variate Observations. Berkeley Symposium on Mathematical Statistics and Probability (1967) 281-297 3. Kohonen, T.: Self-Organizing Maps. Springer, Germany. (1995) 4. Hart, P. E.: The Condensed Nearest Neighbor Rule. IEEE Trans. on Information Theory 14(5) (1967) 515-516 5. Devijver, P. A. & Kittler, J.: Pattern Recognition: A Statistical Approach. Prentice/HallLondon. (1982) 6. Wilson, D. R. & Martinez, T. R.: Reduction Techniques for Instance-Based Learning Algorithms. Machine Learning 38(3) (2000) 257-286 7. Zhao, H. & Lu, B. L.: A Modular k-nearest Neighbor Classification Method for Massively Parallel Text Categorization. In: Zhang, J., He, J.-H., Fu, Y. (eds.): Computational and Information Science: First International Symposium, CIS 2004. Lecture Notes in Computer Science, Vol. 3314. Springer-Verlag, Berlin Heidelberg New York (2004) 867-872 8. Giacinto, G. & Roli, F.: Automatic Design of Multiple Classi-fier Systems by Unsupervised Learning. Proceedings of Machine Learning and Data Mining in Pattern Recognition, First Interna-tional Workshop, MLDM 99, Leipzig, Germany (1999) 131-143 9. Frosyniotis, D., Stafylopatis A. & Likas, A.: A Divide-and-conquer Method for Multi-net Classifiers. Pattern Anal Applic 6 (2003) 32-40 10. Zhao, H.: A Study on Min-max Modular Classifier (in Chinese). Doctoral dissertation of Shanghai Jiao Tong University (2005) 11. Blake, C. L. & Merz, C. J.: UCI Repository of Machine Learning Data-bases [http://www.ics.uci.edu/ mlearn/mlrepository.html]. Irvine, CA: University of California, Department of Information and Com-puter Science (1998) 12. Ratsch, G., Onoda T. & Muller, K.: Soft Margins for AdaBoost. Machine Learning (2000) 1-35 13. Lu, B. L. & Ito, M.: Task Decomposition and Module Combination Based on Class Relations: a Modular Neural Network for Pattern Classification. IEEE Transactions on Neural Networks 10 (1999) 1244-1256 14. Chawla, N. V., Hall, L. O., Bowyer, K. W. & Kegelmeyer, K. P.: Learning Ensembles from Bites: A Scalable and Accurate Approach. Journal of Machine Learning Research 5 (2004) 421-451 15. Yang, Y. & Lu, B. L.: Structure Pruning Strategies for Min-Max Modular Network. In: Wang, J., Liao, X. and Yi Z. (eds.): Advances in Neural Networks C ISNN 2005: Second International Symposium on Neural Networks. Lecture Notes in Computer Science, Vol. 3496. Springer-Verlag, Berlin Heidelberg New York (2005) 646-651 16. Wilson, D. L.: Asymptotic Properties of Nearest Neighbor Rules Using Edited Data. IEEE Transactions on Systems, Man, and Cybernetics 2-3 (1972) 408-421