Learning to assess from pair-wise comparisons 1

Size: px
Start display at page:

Download "Learning to assess from pair-wise comparisons 1"

Transcription

1 Learning to assess from pair-wise comparisons 1 (**) J. Díez (*), J.J. del Coz (**), O. Luaces (**), F. Goyache (*), J. Alonso (**), A. M. Peña (***), and A. Bahamonde (**) (*) SERIDA-CENSYRA-Somió, C/ Camino de los Claveles 604, E Gijón (Asturias), Spain. Centro de Inteligencia Artificial. Universidad de Oviedo at Gijón, Campus de Viesques, E Gijón (Asturias), Spain. (***) Facultad de Ingeniería. Universidad Distrital Francisco José de Caldas, Bogotá, Colombia. jdiez@serida.org, {juanjo, oluaces}@aic.uniovi.es, fgoyache@serida.org, {jalonso, antonio}@aic.uniovi.es Abstract. In this paper we present an algorithm for learning a function able to assess objects. We assume that our teachers can provide a collection of pairwise comparisons but encounter certain difficulties in assigning a number to the qualities of the objects considered. This is a typical situation when dealing with food products, where it is very interesting to have repeatable, reliable mechanisms that are as objective as possible to evaluate quality in order to provide markets with products of a uniform quality. The same problem arises when we are trying to learn user preferences in an information retrieval system or in configuring a complex device. The algorithm is implemented using a growing variant of Kohonen s Self-Organizing Maps (growing neural gas), and is tested with a variety of data sets to demonstrate the capabilities of our approach. 1. Introduction Generally speaking, quality assessment is a complex matter: what we usually need to evaluate are the desirable traits of an object by means of a single number. Frequently though, this number does not strictly reflect an absolute value, but rather the relative quality of the object with respect to others. This is especially true for objects of a biological origin; their quality is dependent on a not always well defined group of multisensorial properties resulting from their chemical composition, the natural structure of the food elements, their interaction and the way in which they are perceived by human senses [13]. This situation becomes even more complex when we consider quality grading of food products from the viewpoint of experts or consumers. Since no detailed grading specifications exist, experts may adopt a quality profile that considerably exceeds that expected by the consumer [2]. The requirements of consumers are usually based on single attributes that characterize primary senses. 1 The research reported in this paper has been supported in part under MCyT and Feder grant TIC

2 Consequently, the literature reflects disagreement between quality assessments obtained through consumer or expert panels [2, 9]. However, the food industry needs to supply markets with uniform quality products to satisfy consumer demands for normalized quality. Furthermore, if possible, food producers would like to know what the objective (chemical and physical) basis of the assessed quality is from the customer viewpoint so as to improve the acceptability of their products. The straightforward way to build computable procedures to assess objects is to collect a set of representative assessment events and then to apply a machine learning algorithm that employs regression like CUBIST [6], M5 [16, 20], SAFE (System to Acquire Functions from Examples) [15] or BETS (Best Examples in Training Sets) [7]. However, our experience with biological objects [9, 10] tells us that the complexity of the assessment task means that the repeatability of human evaluations tends to be low. Hence, the reliability of the training material is poor, despite experts having been trained exhaustively and having accumulated a large, valuable body of knowledge used for assessing [12]. Experts or consumers are perfectly able to prefer one object to another, but usually fail when they are asked to label products with a number. There is a kind of batch effect that often biases the assessment; human assessors try to number the differences in a relative sense, comparing products with the other partners in the batch. Thus, a product surrounded by worse things will probably obtain a higher assessment than if it were presented with better products. However, although we may find unacceptable individual variability in the absolute number obtained to assess the quality of a given product, the relative position obtained in a batch is quite constant. In this paper we present a new approach to learning functions capable of assessing objects starting from reliable training material. Our training sets are formed by pairs of object descriptions, given by continuous attributes, where the first one has been considered worse than the second. The goal is then a function able to quantify the quality of objects as coherently as possible with the pair-wise ordering supplied as the training set. The core idea is to consider each training instance as an indication of a direction where we can find an increase in quality. Thus, the vectorial difference of compared products is interpreted as a kind of coding of the local behavior of the assessment function. In this way, the learning algorithm is a clustering that uses a growing version [8] of Kohonen s Self-Organizing Maps (SOM) [11], where each cell encapsulates a regression rule. After presenting the geometrical motivations of the algorithm followed by the implementation details, we close the paper with a section devoted to presenting the experimental results obtained with our assessment learner. 2. Geometrical motivation of the algorithm Let u and v be vectors describing the features of two objects that our experts compare, resulting in u being worse than v; in symbols, u < v. Then, we seek a function f such that f(u) < f(v). If we assume that f behaves linearly, at least in the surroundings of our vectors, we have to find a vector w such that

3 f w(u) = u w < v w = f w(v) (1) where, for vectors z and t, we represent their inner product by z t. From a geometrical point of view, function f w represents the distance to the hyperplane u w = 0; i.e. the hyperplane of vectors perpendicular to w. If we search for w considering only normalized vectors (i.e. w = 1), the largest difference between f w (u) and f w (v) values is reached when w is the normalized vector in the address of (vu). In fact, f w (v u) = (v u) w v-u w = v-u = (v u) (v u)/ v-u = f (v-u)/ v-u (v u) (2) w u v assessment hyperplane Fig. 1. Given two objects represented by vectors u and v, if u is worse than v, the normal vector in the direction of the difference, w = (v-u)/ v-u, defines a hyperplane, the distance from which is a suitable local assessment function. In the general case we start from a family of comparisons {u i < v i : i = 1,, n } (3) and wish to induce a function f, with local linear behavior, and which hopefully is capable of distinguishing as often as possible that u i is worse than v i, because f(u i ) < f(v i ). The algorithm proposed in this paper uses the geometrical intuition introduced in this section as the basic building block of such a function f. Hence, the main task of the algorithm will be to combine the local guidelines suggested by each comparison supplied in the training set. 3. The algorithm: clustering partial functions In line with the discussions presented in the previous section, the comparison examples of (3) give rise to a set of 2n pairs of vectors as follows: {((v i - u i )/ v i - u i u i ): i = 1,, n} {((v i - u i ) / v i - u i, v i,): i = 1,, n}. (4)

4 If (w, u) is such a pair, we understand it to be a suggestion of a regression rule indicating that the assessment of a vector z is f w (z) if z is in the vicinity of u. (5) Given that f w (z) = z w, we will usually identify w with the linear function f w. Likewise, we will refer to u as the conditions of the rule. For short, we write w u In general, we are pursuing an assessment function f defined by parts of the whole attribute space. In other words, our assessment function will be given by a list of regression rules (w 1 u 1 ); (w 2 u 2 ); ; (w m u m ) (6) that must be evaluated by means of a minimal distance criterion. In symbols, the function f that is finally induced will work as follows for an arbitrary vector z. f(z) = w k z if z - u k z - u j, j = 1,, m (7) A first attempt to define the list of regression rules (6) is to consider the whole set of pairs (w, u) defined in (4), but these rule set must be improved: it is too big and may contain a lot of noise. Therefore, the idea of our learning algorithm (see Algorithm 1) is to cluster similar conditions u, and then to attach a function w according to the functions of the pairs of the same cluster (see Figure 2). To this end, we use a growing version of Kohonen s Self-Organizing Maps (SOM) [11]: growing neural gas (GNG) of Fritzke [8]. This approach has the advantage that we do not need to define a priori configuration parameters like SOM layout dimensions or the radius used throughout the adaptation. Fig. 2. The clusters of partial functions represents, in each node, an environment in the attribute space of the objects to be assessed, drawn in gray in the picture, and a vector pointing in the direction to measure the assessments. In other words, the map represents a set of regression rules to be applied by means of a nearest-distance criterion. The GNG graph starts with two nodes u 1, u 2 representing two points in the domain of the assessment function, in each iteration step a new node is added trying to fit better this space. The number of steps (N) followed by GNG conditions the granularity of the regression rules. By default, N is the number of comparisons divided by 10. Once we have a number of clusters represented by u 1, u 2,,u m, we consider the set of comparisons (t 2 < t 1 ) where each t k is closer to the same u i than to any other u j.

5 These comparisons will be used to compute a local linear approximation of the assessment function in the surroundings of u i. The procedure followed to find a linear function with coefficients w = (a 1,, a d ) is taken from OC1 [14] only slightly modified for this purpose. In fact, what we are looking for is a vector w such that w (t 1 ) > 0 as many times as possible. We can start with w being the average of the normalized differences w = (a 1,, a d ) = Average{(t 1 )/ t 1 t 1 >t 2 & t 1,t 2 cluster(u i )}. (8) Now we try to improve the coefficients a m, one at a time. The key observation is that for each normalized difference (t 1 )/ t 1 =(x 1,, x d ) we have that when x m > 0, is equivalent to w (t 1 ) = Σ(a i * x i : i = 1..d) > 0, (9) a m > -(a 1 x 1 + a 2 x a m-1 x m-1 + a m+1 x m a d x d ) / x m = U (10) or the opposite when x m < 0. When x m = 0, the value of the coefficient a m does not matter. So, for fixed values of all other coefficients, each equation (10) represents a constraint on the values of a m. Therefore, we sort all U values and consider as possible setting for a m the midpoints between each pair of consecutive U s. We select the a m that satisfies the greater number of constraints. Following the procedure of OC1, we iterate this step in order to adjust all the coefficients until no further optimisation can be achieved. If the number of clusters is high, for instance whenever we use the default value for N, the number of training examples divided by 10, then the previous approach inspired in OC1 can be skipped (the results are quite similar). We can simply update the function w attached to a cluster u as the average of the functions w of pairs (w,u ) whose winner node is u. In any case, the regression rules so found need a final improvement process. The idea is that w u may correctly resolve assessments of objects near u. That is, when t 1 >t 2, and both t 1 and t 2 are near u, w was devised for obtaining w (t 1 ) > 0. But w may fail when one of the objects is going to be assessed by another rule w u. To solve these situations we are going to look for adequate slope modifiers a and independent terms b such that the function of the regression rule will now be a(w ) + b u. (11) The procedure followed to find a and b for each regression rule is almost the same that we have just described for adjusting the coefficients a m of each w. The only difference is that now we consider comparisons where only one of the objects is near the condition of the rule to be improved.

6 Function LEARN TO ASSESS COMPARISONS FROM EXAMPLES (LACE) (comparisons set {u i <v i : i=1,,n}, number of steps N) { E = {( vi ui,u i ): i=1,,n} {( vi ui,v i,): i=1,,n}; vi ui vi ui } // To have comparable values in [0,1] Normalize each component of conditions u i and v i in E pairs; // Now, we cluster the conditions of E examples GNG(conditions(E), steps = N); //by default N = E /(2*10) Let (w 1,u 1 ), (w 2,u 2 ),,(w m,u m ) be the nodes of the graph where w i are the average values of the training instances having node i as the nearest one //the next loop can be safety skipped when N is high for each node (w i, u i ) in graph do { //notice that the function w i is an arbitrary value w i = OC1{t 1 : (t 2 < t 1 ) && u i,t 1 j,t 1 u i,t 2 j,t 2 i} } improve relative slopes and independent terms of regression rules; return regression rules; Algorithm 1. The algorithm that learns to assess from pair-wise comparison examples (LACE). 4. Experimental results In order to test the validity of our approach we conducted a number of experiments. The idea is to deal with assessment problems where we know a priori the kind of results that we would like to obtain. To illustrate the way that our algorithm works, we start with a simple problem. Let us consider objects describable by only one continuous attribute x with values in [0, 1], and having as true assessment function the parabola ta(x)= -x(x-1), see Figure (1, a). To build a training set of comparisons E, we generated 3000 pairs (x1,x2) with values in [0,1], and we added to E the pair (x1,x2) if ta(x1) > ta(x2), and we added (x2,x1) otherwise. Our algorithm learned from E the function f drawn in Figure (2, b). Notice that while the actual values of f(x) and ta(x) are quite different, the relative values are almost the same. In fact, building a test set of comparisons using the same procedure followed for E, we only found that the 0.03% of the pairs were erroneously ordered by f. A second package of experiments (see Table 1) was carried out with objects describable by two continuous attributes: x and y. Once an assessment function had been fixed, the objects were randomly generated as 2-dimensional points in the stated rectangles; once we have generated two such objects, they are written in the comparison set, the worse one (according to the corresponding goal assessment function) first. We additionally generated another test set of comparisons, changing the random seed. Both sets had 3000 pairs. The errors reported are the percentage of test pairs that were misplaced by the assessment function learned by our algorithm. These data sets should be easy problems for our learner, and in fact were so, as can be seen in the scores reported in Table (2): However, we can appreciate significantly better scores when the regions with different assessment behavior are separated.

7 (a) (b) Fig. 3. The objects to be assessed are described by x [0, 1]. (a) The true assessment is ta(x) = -x(x-1). (b) The function learned by our algorithm f. Only 1 of the 3000 test pairs is erroneously ordered by f. Table 3. Experiments carried out with goal functions defined by two linear subfunctions with separate domains. The original objects to be assessed were vectors in the rectangles [0, 999] [0,999] in the first two rows, and for the other two [200,299] [0,999], and [700,799] [0,999]. Both training and test sets have 3000 elements. We used only 3 steps to adapt the underlying GNG graph. Goal functions x + 10y x 500 (x,y) = 10x + y x > 500 x + y x 500 (x,y) = x y x > 500 Error f 4.83% f 5.46% x + 10y x [ 200,299] f(x,y) = 10x + y x [ 700,799] 0.23% x + y x [ 200,299] f(x,y) = x y x [ 700,799] 0.20% Finally, we used some publicly available regression datasets in an attempt to deal with almost real-world data. We built training and test sets providing comparisons between the class values of pairs of examples instead of training on their class labels as in the conventional setting; for each example we randomly selected other 10 examples, 8 of them were placed in the training set and the other 2 went to the test set. In order to compare the achievements of LACE, we used two well-known regression learners: M5 [16, 20], and Cubist [6]. We trained M5 and Cubist with the whole dataset, that is considering not only the description of the objects, but the numeric class too. To test what they learned, we compared the values provided for each component of the comparison. The scores so found are reported in Table (4). Let us remark that the comparative reported in Table (4) is not fair for our LACE. The reason is that regression learners have access to the true numerical classes for all test examples, while LACE can only see pairs where there are differences, but without knowing the amount of those differences. As was pointed out in the introduction, in real-world cases we will not have the numeric classes and so we will not able to use M5 or Cubist.

8 Table 4. Error scores of our learner in publicly available regression datasets in addition to the parabola dataset described above. The CPU, Body fat were dowloaded from Cubist URL [6], while Boston housing, and Liver disorders can be found at UCI Repositoty [1]. The number of steps followed by GNG was the default value, i.e., the number of training comparisons divided by 10. Notice that LACE reached only 0.03% errors when N was 3 in the parabola dataset. dataset Cubist M5' LACE CPU 13.16% 11.00% 11.48% Boston Housing 8.99% 9.19% 7.01% Body fat 17.26% 15.48% 11.10% Liver disorders 31.59% 31.30% 14.63% Parabola 0.86% 9.13% 3.93% Average 14.37% 15.22% 9.63% 5. Related work Tesauro tackled a similar problem in [17] for finding a function able to select the most preferable alternative in his famous backgammon player. His proposal was to enforce a symmetric neural network architecture consisting of two separate subnetworks, one for each object in the comparison. In addition, he enforced that both subnetworks have the same weights (only multiplied by -1 in the output layer). However, this restriction in the training mechanism only worked properly with perceptron networks, at least in his application field. Other perceptron approaches are described in [18,19]. In information retrieval, user preferences were modelled by means of preference predicates learned from a set of comparisons [3, 4, 5]. This is a quite different approach since our aim is to obtain a function able to assess grader preferences with a number; for our purposes it is not enough to know which object is preferable. Additionally, once you have a preference predicate, to order a set of objects is a NPhard problem [5] since the transitivity of the learned predicate is not guaranteed at all. 6. Conclusions In this paper, we have presented a new approach to obtaining sound assessment functions of objects. Our approach allows us to make use of a kind of knowledge capable of satisfactorily ranking a set of objects from the best to the worst, but that fails in assessing the goodness of a single object with an absolute number. Assessments carried out in an absolute way are strongly affected by a batch effect in the sense that they tend to number the quality of an object with respect to the other objects in a batch, but not in an absolute sense, as we hope for when we assign a number to quality. This situation is characteristic of biological objects, and especially in the food industry, in which the rules for deciding the degree of quality of a product

9 are not usually well defined, but the ranking of products is quite constant and well accepted on the part of consumers and market operators. From a computational point of view, we have to obtain a float function from training sets without categorical or continuous classes. The problem has been tackled with a growing modification of Kohonen s SOM based on a geometrical intuition of the transformations that should be applied to the training data. The algorithm thus built was tested with both artificial and real-world data in order to show the abilities of the method proposed. The results reflect a very high degree of accuracy. The limitations of our approach, which should be overcome in a future work, have to do with the granularity of the underlying GNG graph that clusters training data. Additionally, we hope that an improvement in the placement of conditions (u) in regression rules (w u) would provide a better performance of solutions with a lower number of steps, see Table 3. References 1. Blake, C., Merz, C. J.: UCI Repository of machine learning databases [ Irvine, CA: University of California, Department of Information and Computer Science (1998) 2. Bollen A.F., Kusabs N.J., Holmes G. and Hall M.A.: Comparison of consumer and producer perceptions of mushroom quality. Proc. Integrated View of Fruit and Vegetable Quality International Multidisciplinary Conference, W.J. Florkowski, S.E. Prussia and R.L. Shewfelt (eds.), Georgia (2002) Branting, K.L., Broos, P.S.: Automated Acquisition of User Preferences. International Journal of Human-Computer Studies, (1997) 46: Branting, K.L.: Active Exploration in Instance-Based Preference Modeling, Proceedings of the Third International Conference on Case-Based Reasoning (ICCBR-99), Germany (1999) 5. Cohen, W.W., Shapire, R.E., Singer, Y.: Learning to order things. Journal of Artificial Intelligence Research (1999) 10, Cubist (2000). Release 1.09, 7. Del Coz, J. J., Luaces, O., Quevedo, J. R., Alonso, J., Ranilla, J., Bahamonde, A.: Self- Organizing Cases to Find Paradigms. Lecture Notes in Computer Sciences, Springer- Verlag, Berlin (1999) Vol. 1606, Fritzke, B.: A growing neural gas network learns topologies, Advances in Neural Information Processing Systems 7, G. Tesauro, D. S. Touretzky and T. K. Leen (eds.), MIT Press, Cambridge MA (1995) Goyache, F., Bahamonde, A. Alonso, J., López, S., Alonso, J., del Coz J.J., Quevedo, J.R., Ranilla, J., Luaces, O., Alvarez, I., Royo, L. and Díez J.: The usefulness of Artificial Intelligence techniques to assess subjective quality of products in the food industry. Trends in Food Science and Technology, in press (2002) 10. Goyache, F., del Coz, J.J., Quevedo, J.R., López, S., Alonso, J., Ranilla, J., Luaces, O., Alvarez, I. and Bahamonde, A.: Using artificial intelligence to design and implement a morphological assessment system in beef cattle. Animal Science (2001), 73: Kohonen, T.: Self-Organizing Maps. Springer Series of Information Science. Springer- Verlag, Berlin (1995) 12. Kusabs N., Bollen F., Trigg L., Holmes G., Inglis S.: Objective measurement of mushroom quality. Proc. New Zealand Institute of Agricultural Science and the New

10 Zealand Society for Horticultural Science Annual Convention, Hawke's Bay, New Zealand (1998) 13. Meilgaard, M., Civille, G.V., Carr, B.T.: Sensory evaluation techniques. CRC Press, Inc., Boca Raton, Florida (1987) 14. Murthy, S. K., Kasif, S., Salzberg, S.: A system for induction of oblique decision trees. Journal of Artificial Intelligence Research, (1994) 2, Quevedo, J.R., Bahamonde, A.: Aprendizaje de Funciones Usando Inducción sobre Clasificaciones Discretas. Proceedings CAEPIA'99 VIII Conferencia de la Asociación Española para la Inteligencia Artificial, Murcia, Spain (1999) Vol. I, Quinlan, J. R.: Learning with continuous classes. Proceedings 5 th Australian Joint Conference on Artificial Intelligence. World Scientific, Singapore, (1992), Tesauro, G.: Connectionist learning of expert preferences by comparison training. In Advances in Neural Information Processing Systems, Proceedings NIPS'88, MIT Press (1989) Utgoff, J. P., Clouse, J.: Two kinds of training information for evaluation function learning. In Proceedings AAAI'91, MIT Press (1991) Utgoff, J.P., Saxema, S.: Learning preference predicate. In Proceedings of the Fourth International Workshop on Machine Learning, Morgan Kaufmann, San Francisco (1987) Wang Y., Witten I.H.: Inducing of Model Trees for Predicting Continuous Classes. Proceedings of European Conference on Machine Learning. Prague, Czech Republic (1997)

ARTIFICIAL NEURAL NETWORKS FOR THE COMPUTATION OF TRAFFIC QUEUES

ARTIFICIAL NEURAL NETWORKS FOR THE COMPUTATION OF TRAFFIC QUEUES ARTIFICIAL NEURAL NETWORKS FOR THE COMPUTATION OF TRAFFIC QUEUES Pedro Hernández-Arauzo, Secundino López-García, Antonio Bahamonde Centro de Inteligencia Artificial. Universidad de Oviedo at Gijón Campus

More information

Using Decision Boundary to Analyze Classifiers

Using Decision Boundary to Analyze Classifiers Using Decision Boundary to Analyze Classifiers Zhiyong Yan Congfu Xu College of Computer Science, Zhejiang University, Hangzhou, China yanzhiyong@zju.edu.cn Abstract In this paper we propose to use decision

More information

Estimating Missing Attribute Values Using Dynamically-Ordered Attribute Trees

Estimating Missing Attribute Values Using Dynamically-Ordered Attribute Trees Estimating Missing Attribute Values Using Dynamically-Ordered Attribute Trees Jing Wang Computer Science Department, The University of Iowa jing-wang-1@uiowa.edu W. Nick Street Management Sciences Department,

More information

Topological Correlation

Topological Correlation Topological Correlation K.A.J. Doherty, R.G. Adams and and N. Davey University of Hertfordshire, Department of Computer Science College Lane, Hatfield, Hertfordshire, UK Abstract. Quantifying the success

More information

Stability Assessment of Electric Power Systems using Growing Neural Gas and Self-Organizing Maps

Stability Assessment of Electric Power Systems using Growing Neural Gas and Self-Organizing Maps Stability Assessment of Electric Power Systems using Growing Gas and Self-Organizing Maps Christian Rehtanz, Carsten Leder University of Dortmund, 44221 Dortmund, Germany Abstract. Liberalized competitive

More information

Representation of 2D objects with a topology preserving network

Representation of 2D objects with a topology preserving network Representation of 2D objects with a topology preserving network Francisco Flórez, Juan Manuel García, José García, Antonio Hernández, Departamento de Tecnología Informática y Computación. Universidad de

More information

Two-step Modified SOM for Parallel Calculation

Two-step Modified SOM for Parallel Calculation Two-step Modified SOM for Parallel Calculation Two-step Modified SOM for Parallel Calculation Petr Gajdoš and Pavel Moravec Petr Gajdoš and Pavel Moravec Department of Computer Science, FEECS, VŠB Technical

More information

Modification of the Growing Neural Gas Algorithm for Cluster Analysis

Modification of the Growing Neural Gas Algorithm for Cluster Analysis Modification of the Growing Neural Gas Algorithm for Cluster Analysis Fernando Canales and Max Chacón Universidad de Santiago de Chile; Depto. de Ingeniería Informática, Avda. Ecuador No 3659 - PoBox 10233;

More information

Processing Missing Values with Self-Organized Maps

Processing Missing Values with Self-Organized Maps Processing Missing Values with Self-Organized Maps David Sommer, Tobias Grimm, Martin Golz University of Applied Sciences Schmalkalden Department of Computer Science D-98574 Schmalkalden, Germany Phone:

More information

C-NBC: Neighborhood-Based Clustering with Constraints

C-NBC: Neighborhood-Based Clustering with Constraints C-NBC: Neighborhood-Based Clustering with Constraints Piotr Lasek Chair of Computer Science, University of Rzeszów ul. Prof. St. Pigonia 1, 35-310 Rzeszów, Poland lasek@ur.edu.pl Abstract. Clustering is

More information

Filter methods for feature selection. A comparative study

Filter methods for feature selection. A comparative study Filter methods for feature selection. A comparative study Noelia Sánchez-Maroño, Amparo Alonso-Betanzos, and María Tombilla-Sanromán University of A Coruña, Department of Computer Science, 15071 A Coruña,

More information

Self-Organizing Maps for cyclic and unbounded graphs

Self-Organizing Maps for cyclic and unbounded graphs Self-Organizing Maps for cyclic and unbounded graphs M. Hagenbuchner 1, A. Sperduti 2, A.C. Tsoi 3 1- University of Wollongong, Wollongong, Australia. 2- University of Padova, Padova, Italy. 3- Hong Kong

More information

Rank Measures for Ordering

Rank Measures for Ordering Rank Measures for Ordering Jin Huang and Charles X. Ling Department of Computer Science The University of Western Ontario London, Ontario, Canada N6A 5B7 email: fjhuang33, clingg@csd.uwo.ca Abstract. Many

More information

Research on Applications of Data Mining in Electronic Commerce. Xiuping YANG 1, a

Research on Applications of Data Mining in Electronic Commerce. Xiuping YANG 1, a International Conference on Education Technology, Management and Humanities Science (ETMHS 2015) Research on Applications of Data Mining in Electronic Commerce Xiuping YANG 1, a 1 Computer Science Department,

More information

Improving A Trajectory Index For Topology Conserving Mapping

Improving A Trajectory Index For Topology Conserving Mapping Proceedings of the 8th WSEAS Int. Conference on Automatic Control, Modeling and Simulation, Prague, Czech Republic, March -4, 006 (pp03-08) Improving A Trajectory Index For Topology Conserving Mapping

More information

A comparative study of local classifiers based on clustering techniques and one-layer neural networks

A comparative study of local classifiers based on clustering techniques and one-layer neural networks A comparative study of local classifiers based on clustering techniques and one-layer neural networks Yuridia Gago-Pallares, Oscar Fontenla-Romero and Amparo Alonso-Betanzos University of A Coruña, Department

More information

The Role of Biomedical Dataset in Classification

The Role of Biomedical Dataset in Classification The Role of Biomedical Dataset in Classification Ajay Kumar Tanwani and Muddassar Farooq Next Generation Intelligent Networks Research Center (nexgin RC) National University of Computer & Emerging Sciences

More information

An Empirical Study of Hoeffding Racing for Model Selection in k-nearest Neighbor Classification

An Empirical Study of Hoeffding Racing for Model Selection in k-nearest Neighbor Classification An Empirical Study of Hoeffding Racing for Model Selection in k-nearest Neighbor Classification Flora Yu-Hui Yeh and Marcus Gallagher School of Information Technology and Electrical Engineering University

More information

Improving Classifier Performance by Imputing Missing Values using Discretization Method

Improving Classifier Performance by Imputing Missing Values using Discretization Method Improving Classifier Performance by Imputing Missing Values using Discretization Method E. CHANDRA BLESSIE Assistant Professor, Department of Computer Science, D.J.Academy for Managerial Excellence, Coimbatore,

More information

Recent Progress on RAIL: Automating Clustering and Comparison of Different Road Classification Techniques on High Resolution Remotely Sensed Imagery

Recent Progress on RAIL: Automating Clustering and Comparison of Different Road Classification Techniques on High Resolution Remotely Sensed Imagery Recent Progress on RAIL: Automating Clustering and Comparison of Different Road Classification Techniques on High Resolution Remotely Sensed Imagery Annie Chen ANNIEC@CSE.UNSW.EDU.AU Gary Donovan GARYD@CSE.UNSW.EDU.AU

More information

A Lazy Approach for Machine Learning Algorithms

A Lazy Approach for Machine Learning Algorithms A Lazy Approach for Machine Learning Algorithms Inés M. Galván, José M. Valls, Nicolas Lecomte and Pedro Isasi Abstract Most machine learning algorithms are eager methods in the sense that a model is generated

More information

FLEXIBLE AND OPTIMAL M5 MODEL TREES WITH APPLICATIONS TO FLOW PREDICTIONS

FLEXIBLE AND OPTIMAL M5 MODEL TREES WITH APPLICATIONS TO FLOW PREDICTIONS 6 th International Conference on Hydroinformatics - Liong, Phoon & Babovic (eds) 2004 World Scientific Publishing Company, ISBN 981-238-787-0 FLEXIBLE AND OPTIMAL M5 MODEL TREES WITH APPLICATIONS TO FLOW

More information

Comparing Univariate and Multivariate Decision Trees *

Comparing Univariate and Multivariate Decision Trees * Comparing Univariate and Multivariate Decision Trees * Olcay Taner Yıldız, Ethem Alpaydın Department of Computer Engineering Boğaziçi University, 80815 İstanbul Turkey yildizol@cmpe.boun.edu.tr, alpaydin@boun.edu.tr

More information

Induction of Multivariate Decision Trees by Using Dipolar Criteria

Induction of Multivariate Decision Trees by Using Dipolar Criteria Induction of Multivariate Decision Trees by Using Dipolar Criteria Leon Bobrowski 1,2 and Marek Krȩtowski 1 1 Institute of Computer Science, Technical University of Bia lystok, Poland 2 Institute of Biocybernetics

More information

Mining High Order Decision Rules

Mining High Order Decision Rules Mining High Order Decision Rules Y.Y. Yao Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 e-mail: yyao@cs.uregina.ca Abstract. We introduce the notion of high

More information

Automatic Group-Outlier Detection

Automatic Group-Outlier Detection Automatic Group-Outlier Detection Amine Chaibi and Mustapha Lebbah and Hanane Azzag LIPN-UMR 7030 Université Paris 13 - CNRS 99, av. J-B Clément - F-93430 Villetaneuse {firstname.secondname}@lipn.univ-paris13.fr

More information

Improving Tree-Based Classification Rules Using a Particle Swarm Optimization

Improving Tree-Based Classification Rules Using a Particle Swarm Optimization Improving Tree-Based Classification Rules Using a Particle Swarm Optimization Chi-Hyuck Jun *, Yun-Ju Cho, and Hyeseon Lee Department of Industrial and Management Engineering Pohang University of Science

More information

Efficient Pairwise Classification

Efficient Pairwise Classification Efficient Pairwise Classification Sang-Hyeun Park and Johannes Fürnkranz TU Darmstadt, Knowledge Engineering Group, D-64289 Darmstadt, Germany Abstract. Pairwise classification is a class binarization

More information

Statistical dependence measure for feature selection in microarray datasets

Statistical dependence measure for feature selection in microarray datasets Statistical dependence measure for feature selection in microarray datasets Verónica Bolón-Canedo 1, Sohan Seth 2, Noelia Sánchez-Maroño 1, Amparo Alonso-Betanzos 1 and José C. Príncipe 2 1- Department

More information

An Information-Theoretic Approach to the Prepruning of Classification Rules

An Information-Theoretic Approach to the Prepruning of Classification Rules An Information-Theoretic Approach to the Prepruning of Classification Rules Max Bramer University of Portsmouth, Portsmouth, UK Abstract: Keywords: The automatic induction of classification rules from

More information

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used. 1 4.12 Generalization In back-propagation learning, as many training examples as possible are typically used. It is hoped that the network so designed generalizes well. A network generalizes well when

More information

Model-based segmentation and recognition from range data

Model-based segmentation and recognition from range data Model-based segmentation and recognition from range data Jan Boehm Institute for Photogrammetry Universität Stuttgart Germany Keywords: range image, segmentation, object recognition, CAD ABSTRACT This

More information

WEIGHTED K NEAREST NEIGHBOR CLASSIFICATION ON FEATURE PROJECTIONS 1

WEIGHTED K NEAREST NEIGHBOR CLASSIFICATION ON FEATURE PROJECTIONS 1 WEIGHTED K NEAREST NEIGHBOR CLASSIFICATION ON FEATURE PROJECTIONS 1 H. Altay Güvenir and Aynur Akkuş Department of Computer Engineering and Information Science Bilkent University, 06533, Ankara, Turkey

More information

Supervised Hybrid SOM-NG Algorithm

Supervised Hybrid SOM-NG Algorithm ADVCOMP : The Fifth International Conference on Advanced Engineering Computing and Applications in Sciences Supervised Hybrid -NG Algorithm Mario J. Crespo-Ramos, Iván Machón-González, Hilario López-García

More information

Formal Model. Figure 1: The target concept T is a subset of the concept S = [0, 1]. The search agent needs to search S for a point in T.

Formal Model. Figure 1: The target concept T is a subset of the concept S = [0, 1]. The search agent needs to search S for a point in T. Although this paper analyzes shaping with respect to its benefits on search problems, the reader should recognize that shaping is often intimately related to reinforcement learning. The objective in reinforcement

More information

Improving the Random Forest Algorithm by Randomly Varying the Size of the Bootstrap Samples for Low Dimensional Data Sets

Improving the Random Forest Algorithm by Randomly Varying the Size of the Bootstrap Samples for Low Dimensional Data Sets Improving the Random Forest Algorithm by Randomly Varying the Size of the Bootstrap Samples for Low Dimensional Data Sets Md Nasim Adnan and Md Zahidul Islam Centre for Research in Complex Systems (CRiCS)

More information

Toward Part-based Document Image Decoding

Toward Part-based Document Image Decoding 2012 10th IAPR International Workshop on Document Analysis Systems Toward Part-based Document Image Decoding Wang Song, Seiichi Uchida Kyushu University, Fukuoka, Japan wangsong@human.ait.kyushu-u.ac.jp,

More information

DESIGN OF KOHONEN SELF-ORGANIZING MAP WITH REDUCED STRUCTURE

DESIGN OF KOHONEN SELF-ORGANIZING MAP WITH REDUCED STRUCTURE DESIGN OF KOHONEN SELF-ORGANIZING MAP WITH REDUCED STRUCTURE S. Kajan, M. Lajtman Institute of Control and Industrial Informatics, Faculty of Electrical Engineering and Information Technology, Slovak University

More information

entire search space constituting coefficient sets. The brute force approach performs three passes through the search space, with each run the se

entire search space constituting coefficient sets. The brute force approach performs three passes through the search space, with each run the se Evolving Simulation Modeling: Calibrating SLEUTH Using a Genetic Algorithm M. D. Clarke-Lauer 1 and Keith. C. Clarke 2 1 California State University, Sacramento, 625 Woodside Sierra #2, Sacramento, CA,

More information

Online labelling strategies for growing neural gas

Online labelling strategies for growing neural gas Online labelling strategies for growing neural gas Oliver Beyer and Philipp Cimiano Semantic Computing Group, CITEC, Bielefeld University, obeyer@cit-ec.uni-bielefeld.de http://www.sc.cit-ec.uni-bielefeld.de

More information

USING IMAGES PATTERN RECOGNITION AND NEURAL NETWORKS FOR COATING QUALITY ASSESSMENT Image processing for quality assessment

USING IMAGES PATTERN RECOGNITION AND NEURAL NETWORKS FOR COATING QUALITY ASSESSMENT Image processing for quality assessment USING IMAGES PATTERN RECOGNITION AND NEURAL NETWORKS FOR COATING QUALITY ASSESSMENT Image processing for quality assessment L.-M. CHANG and Y.A. ABDELRAZIG School of Civil Engineering, Purdue University,

More information

Constructively Learning a Near-Minimal Neural Network Architecture

Constructively Learning a Near-Minimal Neural Network Architecture Constructively Learning a Near-Minimal Neural Network Architecture Justin Fletcher and Zoran ObradoviC Abetract- Rather than iteratively manually examining a variety of pre-specified architectures, a constructive

More information

ORT EP R RCH A ESE R P A IDI! " #$$% &' (# $!"

ORT EP R RCH A ESE R P A IDI!  #$$% &' (# $! R E S E A R C H R E P O R T IDIAP A Parallel Mixture of SVMs for Very Large Scale Problems Ronan Collobert a b Yoshua Bengio b IDIAP RR 01-12 April 26, 2002 Samy Bengio a published in Neural Computation,

More information

Genetic Tuning for Improving Wang and Mendel s Fuzzy Database

Genetic Tuning for Improving Wang and Mendel s Fuzzy Database Proceedings of the 2009 IEEE International Conference on Systems, Man, and Cybernetics San Antonio, TX, USA - October 2009 Genetic Tuning for Improving Wang and Mendel s Fuzzy Database E. R. R. Kato, O.

More information

Figure (5) Kohonen Self-Organized Map

Figure (5) Kohonen Self-Organized Map 2- KOHONEN SELF-ORGANIZING MAPS (SOM) - The self-organizing neural networks assume a topological structure among the cluster units. - There are m cluster units, arranged in a one- or two-dimensional array;

More information

Swarm Based Fuzzy Clustering with Partition Validity

Swarm Based Fuzzy Clustering with Partition Validity Swarm Based Fuzzy Clustering with Partition Validity Lawrence O. Hall and Parag M. Kanade Computer Science & Engineering Dept University of South Florida, Tampa FL 33620 @csee.usf.edu Abstract

More information

Rule extraction from support vector machines

Rule extraction from support vector machines Rule extraction from support vector machines Haydemar Núñez 1,3 Cecilio Angulo 1,2 Andreu Català 1,2 1 Dept. of Systems Engineering, Polytechnical University of Catalonia Avda. Victor Balaguer s/n E-08800

More information

A Study on Clustering Method by Self-Organizing Map and Information Criteria

A Study on Clustering Method by Self-Organizing Map and Information Criteria A Study on Clustering Method by Self-Organizing Map and Information Criteria Satoru Kato, Tadashi Horiuchi,andYoshioItoh Matsue College of Technology, 4-4 Nishi-ikuma, Matsue, Shimane 90-88, JAPAN, kato@matsue-ct.ac.jp

More information

Controlling the spread of dynamic self-organising maps

Controlling the spread of dynamic self-organising maps Neural Comput & Applic (2004) 13: 168 174 DOI 10.1007/s00521-004-0419-y ORIGINAL ARTICLE L. D. Alahakoon Controlling the spread of dynamic self-organising maps Received: 7 April 2004 / Accepted: 20 April

More information

2 cubist.default. Fit a Cubist model

2 cubist.default. Fit a Cubist model Package Cubist August 29, 2013 Type Package Title Rule- and Instance-Based Regression Modeling Version 0.0.13 Date 2013-01-28 Author Max Kuhn, Steve Weston, Chris Keefer, Nathan Coulter. C code for Cubist

More information

Improvements to the SMO Algorithm for SVM Regression

Improvements to the SMO Algorithm for SVM Regression 1188 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 5, SEPTEMBER 2000 Improvements to the SMO Algorithm for SVM Regression S. K. Shevade, S. S. Keerthi, C. Bhattacharyya, K. R. K. Murthy Abstract This

More information

Weighting and selection of features.

Weighting and selection of features. Intelligent Information Systems VIII Proceedings of the Workshop held in Ustroń, Poland, June 14-18, 1999 Weighting and selection of features. Włodzisław Duch and Karol Grudziński Department of Computer

More information

Graph Matching: Fast Candidate Elimination Using Machine Learning Techniques

Graph Matching: Fast Candidate Elimination Using Machine Learning Techniques Graph Matching: Fast Candidate Elimination Using Machine Learning Techniques M. Lazarescu 1,2, H. Bunke 1, and S. Venkatesh 2 1 Computer Science Department, University of Bern, Switzerland 2 School of

More information

Designing Radial Basis Neural Networks using a Distributed Architecture

Designing Radial Basis Neural Networks using a Distributed Architecture Designing Radial Basis Neural Networks using a Distributed Architecture J.M. Valls, A. Leal, I.M. Galván and J.M. Molina Computer Science Department Carlos III University of Madrid Avenida de la Universidad,

More information

Supervised Variable Clustering for Classification of NIR Spectra

Supervised Variable Clustering for Classification of NIR Spectra Supervised Variable Clustering for Classification of NIR Spectra Catherine Krier *, Damien François 2, Fabrice Rossi 3, Michel Verleysen, Université catholique de Louvain, Machine Learning Group, place

More information

Investigation of Alternative Strategies and Quality Measures for Controlling the Growth Process of the Growing Hierarchical Self-Organizing Map

Investigation of Alternative Strategies and Quality Measures for Controlling the Growth Process of the Growing Hierarchical Self-Organizing Map Investigation of Alternative Strategies and Quality Measures for Controlling the Growth Process of the Growing Hierarchical Self-Organizing Map Michael Dittenbach ispaces Group ecommerce Competence Center

More information

Feature-weighted k-nearest Neighbor Classifier

Feature-weighted k-nearest Neighbor Classifier Proceedings of the 27 IEEE Symposium on Foundations of Computational Intelligence (FOCI 27) Feature-weighted k-nearest Neighbor Classifier Diego P. Vivencio vivencio@comp.uf scar.br Estevam R. Hruschka

More information

3D Hand Pose Estimation with Neural Networks

3D Hand Pose Estimation with Neural Networks 3D Hand Pose Estimation with Neural Networks Jose Antonio Serra 1, Jose Garcia-Rodriguez 1, Sergios Orts 1, Juan Manuel Garcia- Chamizo 1, Anastassia Angelopoulou 2, Alexandra Psarou 2, Markos Mentzelopoulos

More information

Perceptron-Based Oblique Tree (P-BOT)

Perceptron-Based Oblique Tree (P-BOT) Perceptron-Based Oblique Tree (P-BOT) Ben Axelrod Stephen Campos John Envarli G.I.T. G.I.T. G.I.T. baxelrod@cc.gatech sjcampos@cc.gatech envarli@cc.gatech Abstract Decision trees are simple and fast data

More information

Selection of n in K-Means Algorithm

Selection of n in K-Means Algorithm International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 4, Number 6 (2014), pp. 577-582 International Research Publications House http://www. irphouse.com Selection of n in

More information

Classification Using Unstructured Rules and Ant Colony Optimization

Classification Using Unstructured Rules and Ant Colony Optimization Classification Using Unstructured Rules and Ant Colony Optimization Negar Zakeri Nejad, Amir H. Bakhtiary, and Morteza Analoui Abstract In this paper a new method based on the algorithm is proposed to

More information

Simplifying OCR Neural Networks with Oracle Learning

Simplifying OCR Neural Networks with Oracle Learning SCIMA 2003 - International Workshop on Soft Computing Techniques in Instrumentation, Measurement and Related Applications Provo, Utah, USA, 17 May 2003 Simplifying OCR Neural Networks with Oracle Learning

More information

Growing Neural Gas A Parallel Approach

Growing Neural Gas A Parallel Approach Growing Neural Gas A Parallel Approach Lukáš Vojáček 1 and JiříDvorský 2 1 IT4Innovations Centre of Excellence Ostrava, Czech Republic lukas.vojacek@vsb.cz 2 Department of Computer Science, VŠB Technical

More information

efmea RAISING EFFICIENCY OF FMEA BY MATRIX-BASED FUNCTION AND FAILURE NETWORKS

efmea RAISING EFFICIENCY OF FMEA BY MATRIX-BASED FUNCTION AND FAILURE NETWORKS efmea RAISING EFFICIENCY OF FMEA BY MATRIX-BASED FUNCTION AND FAILURE NETWORKS Maik Maurer Technische Universität München, Product Development, Boltzmannstr. 15, 85748 Garching, Germany. Email: maik.maurer@pe.mw.tum.de

More information

Package Cubist. December 2, 2017

Package Cubist. December 2, 2017 Type Package Package Cubist December 2, 2017 Title Rule- And Instance-Based Regression Modeling Version 0.2.1 Date 2017-12-01 Maintainer Max Kuhn Description Regression modeling using

More information

Extract an Essential Skeleton of a Character as a Graph from a Character Image

Extract an Essential Skeleton of a Character as a Graph from a Character Image Extract an Essential Skeleton of a Character as a Graph from a Character Image Kazuhisa Fujita University of Electro-Communications 1-5-1 Chofugaoka, Chofu, Tokyo, 182-8585 Japan k-z@nerve.pc.uec.ac.jp

More information

BRACE: A Paradigm For the Discretization of Continuously Valued Data

BRACE: A Paradigm For the Discretization of Continuously Valued Data Proceedings of the Seventh Florida Artificial Intelligence Research Symposium, pp. 7-2, 994 BRACE: A Paradigm For the Discretization of Continuously Valued Data Dan Ventura Tony R. Martinez Computer Science

More information

A Comparative Study of Data Mining Process Models (KDD, CRISP-DM and SEMMA)

A Comparative Study of Data Mining Process Models (KDD, CRISP-DM and SEMMA) International Journal of Innovation and Scientific Research ISSN 2351-8014 Vol. 12 No. 1 Nov. 2014, pp. 217-222 2014 Innovative Space of Scientific Research Journals http://www.ijisr.issr-journals.org/

More information

Efficient Pairwise Classification

Efficient Pairwise Classification Efficient Pairwise Classification Sang-Hyeun Park and Johannes Fürnkranz TU Darmstadt, Knowledge Engineering Group, D-64289 Darmstadt, Germany {park,juffi}@ke.informatik.tu-darmstadt.de Abstract. Pairwise

More information

Correlation and Relaxation Labelling---An Experimental Investigation on Fast Algorithms

Correlation and Relaxation Labelling---An Experimental Investigation on Fast Algorithms Correlation and Relaxation Labelling---An Experimental Investigation on Fast Algorithms Q. X. Wu*, S. J. McNeill, & D. Pairman Landcare Research New Zealand PO Box 38491, Wellington Mail Centre New Zealand

More information

Estimating Feature Discriminant Power in Decision Tree Classifiers*

Estimating Feature Discriminant Power in Decision Tree Classifiers* Estimating Feature Discriminant Power in Decision Tree Classifiers* I. Gracia 1, F. Pla 1, F. J. Ferri 2 and P. Garcia 1 1 Departament d'inform~tica. Universitat Jaume I Campus Penyeta Roja, 12071 Castell6.

More information

Accelerating Improvement of Fuzzy Rules Induction with Artificial Immune Systems

Accelerating Improvement of Fuzzy Rules Induction with Artificial Immune Systems Accelerating Improvement of Fuzzy Rules Induction with Artificial Immune Systems EDWARD MĘŻYK, OLGIERD UNOLD Institute of Computer Engineering, Control and Robotics Wroclaw University of Technology Wyb.

More information

Feature weighting using particle swarm optimization for learning vector quantization classifier

Feature weighting using particle swarm optimization for learning vector quantization classifier Journal of Physics: Conference Series PAPER OPEN ACCESS Feature weighting using particle swarm optimization for learning vector quantization classifier To cite this article: A Dongoran et al 2018 J. Phys.:

More information

Efficient SQL-Querying Method for Data Mining in Large Data Bases

Efficient SQL-Querying Method for Data Mining in Large Data Bases Efficient SQL-Querying Method for Data Mining in Large Data Bases Nguyen Hung Son Institute of Mathematics Warsaw University Banacha 2, 02095, Warsaw, Poland Abstract Data mining can be understood as a

More information

Table of Contents. Recognition of Facial Gestures... 1 Attila Fazekas

Table of Contents. Recognition of Facial Gestures... 1 Attila Fazekas Table of Contents Recognition of Facial Gestures...................................... 1 Attila Fazekas II Recognition of Facial Gestures Attila Fazekas University of Debrecen, Institute of Informatics

More information

The Projected Dip-means Clustering Algorithm

The Projected Dip-means Clustering Algorithm Theofilos Chamalis Department of Computer Science & Engineering University of Ioannina GR 45110, Ioannina, Greece thchama@cs.uoi.gr ABSTRACT One of the major research issues in data clustering concerns

More information

SIMULTANEOUS COMPUTATION OF MODEL ORDER AND PARAMETER ESTIMATION FOR ARX MODEL BASED ON MULTI- SWARM PARTICLE SWARM OPTIMIZATION

SIMULTANEOUS COMPUTATION OF MODEL ORDER AND PARAMETER ESTIMATION FOR ARX MODEL BASED ON MULTI- SWARM PARTICLE SWARM OPTIMIZATION SIMULTANEOUS COMPUTATION OF MODEL ORDER AND PARAMETER ESTIMATION FOR ARX MODEL BASED ON MULTI- SWARM PARTICLE SWARM OPTIMIZATION Kamil Zakwan Mohd Azmi, Zuwairie Ibrahim and Dwi Pebrianti Faculty of Electrical

More information

A Wrapper for Reweighting Training Instances for Handling Imbalanced Data Sets

A Wrapper for Reweighting Training Instances for Handling Imbalanced Data Sets A Wrapper for Reweighting Training Instances for Handling Imbalanced Data Sets M. Karagiannopoulos, D. Anyfantis, S. Kotsiantis and P. Pintelas Educational Software Development Laboratory Department of

More information

Lecture #11: The Perceptron

Lecture #11: The Perceptron Lecture #11: The Perceptron Mat Kallada STAT2450 - Introduction to Data Mining Outline for Today Welcome back! Assignment 3 The Perceptron Learning Method Perceptron Learning Rule Assignment 3 Will be

More information

LOGICAL OPERATOR USAGE IN STRUCTURAL MODELLING

LOGICAL OPERATOR USAGE IN STRUCTURAL MODELLING LOGICAL OPERATOR USAGE IN STRUCTURAL MODELLING Ieva Zeltmate (a) (a) Riga Technical University, Faculty of Computer Science and Information Technology Department of System Theory and Design ieva.zeltmate@gmail.com

More information

A novel firing rule for training Kohonen selforganising

A novel firing rule for training Kohonen selforganising A novel firing rule for training Kohonen selforganising maps D. T. Pham & A. B. Chan Manufacturing Engineering Centre, School of Engineering, University of Wales Cardiff, P.O. Box 688, Queen's Buildings,

More information

K-Means Clustering With Initial Centroids Based On Difference Operator

K-Means Clustering With Initial Centroids Based On Difference Operator K-Means Clustering With Initial Centroids Based On Difference Operator Satish Chaurasiya 1, Dr.Ratish Agrawal 2 M.Tech Student, School of Information and Technology, R.G.P.V, Bhopal, India Assistant Professor,

More information

Leave-One-Out Support Vector Machines

Leave-One-Out Support Vector Machines Leave-One-Out Support Vector Machines Jason Weston Department of Computer Science Royal Holloway, University of London, Egham Hill, Egham, Surrey, TW20 OEX, UK. Abstract We present a new learning algorithm

More information

Query-Sensitive Similarity Measure for Content-Based Image Retrieval

Query-Sensitive Similarity Measure for Content-Based Image Retrieval Query-Sensitive Similarity Measure for Content-Based Image Retrieval Zhi-Hua Zhou Hong-Bin Dai National Laboratory for Novel Software Technology Nanjing University, Nanjing 2193, China {zhouzh, daihb}@lamda.nju.edu.cn

More information

Inducer: a Rule Induction Workbench for Data Mining

Inducer: a Rule Induction Workbench for Data Mining Inducer: a Rule Induction Workbench for Data Mining Max Bramer Faculty of Technology University of Portsmouth Portsmouth, UK Email: Max.Bramer@port.ac.uk Fax: +44-2392-843030 Abstract One of the key technologies

More information

Cluster homogeneity as a semi-supervised principle for feature selection using mutual information

Cluster homogeneity as a semi-supervised principle for feature selection using mutual information Cluster homogeneity as a semi-supervised principle for feature selection using mutual information Frederico Coelho 1 and Antonio Padua Braga 1 andmichelverleysen 2 1- Universidade Federal de Minas Gerais

More information

Data mining with Support Vector Machine

Data mining with Support Vector Machine Data mining with Support Vector Machine Ms. Arti Patle IES, IPS Academy Indore (M.P.) artipatle@gmail.com Mr. Deepak Singh Chouhan IES, IPS Academy Indore (M.P.) deepak.schouhan@yahoo.com Abstract: Machine

More information

Combining Deep Learning and Preference Learning for Object Tracking

Combining Deep Learning and Preference Learning for Object Tracking Combining Deep Learning and Preference Learning for Object Tracking Shuchao Pang 1,2, Juan José del Coz 2, Zhezhou Yu 1, Oscar Luaces 2, and Jorge Díez 2 1 Coll. Computer Science and Technology, Jilin

More information

Logical Decision Rules: Teaching C4.5 to Speak Prolog

Logical Decision Rules: Teaching C4.5 to Speak Prolog Logical Decision Rules: Teaching C4.5 to Speak Prolog Kamran Karimi and Howard J. Hamilton Department of Computer Science University of Regina Regina, Saskatchewan Canada S4S 0A2 {karimi,hamilton}@cs.uregina.ca

More information

A Modular Reduction Method for k-nn Algorithm with Self-recombination Learning

A Modular Reduction Method for k-nn Algorithm with Self-recombination Learning A Modular Reduction Method for k-nn Algorithm with Self-recombination Learning Hai Zhao and Bao-Liang Lu Department of Computer Science and Engineering, Shanghai Jiao Tong University, 800 Dong Chuan Rd.,

More information

This is a repository copy of A Rule Chaining Architecture Using a Correlation Matrix Memory.

This is a repository copy of A Rule Chaining Architecture Using a Correlation Matrix Memory. This is a repository copy of A Rule Chaining Architecture Using a Correlation Matrix Memory. White Rose Research Online URL for this paper: http://eprints.whiterose.ac.uk/88231/ Version: Submitted Version

More information

Improving the Hopfield Network through Beam Search

Improving the Hopfield Network through Beam Search Brigham Young University BYU ScholarsArchive All Faculty Publications 2001-07-19 Improving the Hopfield Network through Beam Search Tony R. Martinez martinez@cs.byu.edu Xinchuan Zeng Follow this and additional

More information

Cost-sensitive C4.5 with post-pruning and competition

Cost-sensitive C4.5 with post-pruning and competition Cost-sensitive C4.5 with post-pruning and competition Zilong Xu, Fan Min, William Zhu Lab of Granular Computing, Zhangzhou Normal University, Zhangzhou 363, China Abstract Decision tree is an effective

More information

HALF&HALF BAGGING AND HARD BOUNDARY POINTS. Leo Breiman Statistics Department University of California Berkeley, CA

HALF&HALF BAGGING AND HARD BOUNDARY POINTS. Leo Breiman Statistics Department University of California Berkeley, CA 1 HALF&HALF BAGGING AND HARD BOUNDARY POINTS Leo Breiman Statistics Department University of California Berkeley, CA 94720 leo@stat.berkeley.edu Technical Report 534 Statistics Department September 1998

More information

Seismic regionalization based on an artificial neural network

Seismic regionalization based on an artificial neural network Seismic regionalization based on an artificial neural network *Jaime García-Pérez 1) and René Riaño 2) 1), 2) Instituto de Ingeniería, UNAM, CU, Coyoacán, México D.F., 014510, Mexico 1) jgap@pumas.ii.unam.mx

More information

A Parallel Evolutionary Algorithm for Discovery of Decision Rules

A Parallel Evolutionary Algorithm for Discovery of Decision Rules A Parallel Evolutionary Algorithm for Discovery of Decision Rules Wojciech Kwedlo Faculty of Computer Science Technical University of Bia lystok Wiejska 45a, 15-351 Bia lystok, Poland wkwedlo@ii.pb.bialystok.pl

More information

Towards Automatic Recognition of Fonts using Genetic Approach

Towards Automatic Recognition of Fonts using Genetic Approach Towards Automatic Recognition of Fonts using Genetic Approach M. SARFRAZ Department of Information and Computer Science King Fahd University of Petroleum and Minerals KFUPM # 1510, Dhahran 31261, Saudi

More information

LOAD BALANCING IN MOBILE INTELLIGENT AGENTS FRAMEWORK USING DATA MINING CLASSIFICATION TECHNIQUES

LOAD BALANCING IN MOBILE INTELLIGENT AGENTS FRAMEWORK USING DATA MINING CLASSIFICATION TECHNIQUES 8 th International Conference on DEVELOPMENT AND APPLICATION SYSTEMS S u c e a v a, R o m a n i a, M a y 25 27, 2 0 0 6 LOAD BALANCING IN MOBILE INTELLIGENT AGENTS FRAMEWORK USING DATA MINING CLASSIFICATION

More information

Estimation of Distribution Algorithm Based on Mixture

Estimation of Distribution Algorithm Based on Mixture Estimation of Distribution Algorithm Based on Mixture Qingfu Zhang, Jianyong Sun, Edward Tsang, and John Ford Department of Computer Science, University of Essex CO4 3SQ, Colchester, Essex, U.K. 8th May,

More information

Texture Classification by Combining Local Binary Pattern Features and a Self-Organizing Map

Texture Classification by Combining Local Binary Pattern Features and a Self-Organizing Map Texture Classification by Combining Local Binary Pattern Features and a Self-Organizing Map Markus Turtinen, Topi Mäenpää, and Matti Pietikäinen Machine Vision Group, P.O.Box 4500, FIN-90014 University

More information