Constructive Induction Using Non-algebraic Feature Representation

Size: px
Start display at page:

Download "Constructive Induction Using Non-algebraic Feature Representation"

Transcription

1 Constructive Induction Using Non-algebraic Feature Representation Leila S. Shafti, Eduardo Pérez EPS, Universidad Autónoma de Madrid, Ctra. de Colmenar Viejo, Km 15, E 28049, Madrid, Spain {Leila.Shafti, Eduardo.perez}@ii.uam.es Abstract Learning hard concepts in spite of complex interaction among makes constructive induction necessary. Most constructive induction methods apply a greedy search for constructing new features. The search space of hard concepts with complex interaction among has high variation. Therefore, a greedy constructive induction method falls in local optima when searching this space. A global search such as genetic algorithms is more convenient for hard concepts than a greedy local search. Existing constructive induction methods based on genetic algorithms still suffer from some deficiencies because of their overly restricted representation language, which in turn, defines search space. In this paper we explain how the search space can be decomposed into two spaces and we present a new genetic algorithm constructive induction method based on a non-algebraic representation of features. Experiments show that our method outperforms existing constructive induction methods. Key Words Machine Learning, Constructive Induction, Feature Selection and Construction, Genetic Algorithms 1. Introduction Hard concepts with complex interaction among are difficult to be learned by machine learning algorithms based on similarity. These algorithms assume that cases belonging to the same class are located close to each other. But for hard concepts with complex interaction, each class is scattered through the space due to low-level representation [1], [2]. Interaction means the relation between one attribute and the target concept depends on another attribute. Such problem has been seen in realworld domains such as protein secondary structure [3]. Constructive induction (CI) methods have been introduced to ease the attribute interaction problem. Their goal is to automatically transform the original representation space of hard concepts into a new one where the regularity is more apparent [4], [5]. This goal is achieved by constructing features from the given set of to abstract the relation (interaction) among several to a single new attribute. Most existing CI methods are based on greedy search. These methods have addressed some problems. However, they still have an important limitation; they suffer from the local optima problem. When the concept is complex because of the interaction among, the search space for constructing new features has more variation. Because of high variation, the CI method needs a more global search strategy such as Genetic Algorithms [6] to be able to skip local optima and find the global optima solutions. Genetic Algorithms (GA) are a kind of multidirectional parallel search, and more likely to be successful in searching through the intractable and complicated search space [7]. There are only a few CI methods that use genetic search strategy for constructing new features. Among these methods are GCI [8], GPCI [9], Gabret [10], and the hybrid method of Ritthoff et al [11]. Their partial success in constructing useful features indicates the effectiveness of genetic-based search for CI. However, these methods still have some limitations and deficiencies as will be seen later. The most important ones are their consideration of the search space and the language they apply for representing constructed features. This paper presents DCI, a new CI method based on GA. We begin by reviewing some genetic-based CI methods. Then, we analyze the search space and search method, and the representation language, which are crucial for converging to an optimal solution. Considering these we design a CI method based on GA and present and discuss the results of our initial experiments. 2. Exploring the Search Space A CI method aims to construct features that highlight the interactions. For achieving this aim, it has to look for subsets of interacting and functions over these subsets that capture interactions. Existing genetic-based CI methods, GCI, and GPCI, search the space (of size 2 2N for N input in Boolean domain) of all functions that can be defined over all subsets of. This enormous space has lots of local optima and is more difficult to be explored. To ease the searching task, we propose to decompose the

2 The function that represents interactions Figure 1. The decomposed search space Subset of interacting {x 1,..., x n } Ø search space into two spaces: S the space of all subsets of, and F Si the space of all functions defined over a subset of S i (Figure 1). The decomposition of the search space into two spaces allows a specific method to be adjusted for each search space. This strategy divides the main goal of CI into two easier sub-goals: finding the subset of interacting (S-Search) and looking for a function that represents the interaction among in a given subset (F Si -Search). There are few methods that consider the decomposition of search space. Among these methods are MRP [12], hybrid method of Ritthoff et al [11] and Gabret [10]. These methods apply different search strategies for each space. MRP uses a greedy hill climbing strategy for selecting, starting from set of all and eliminating one attribute at a time. For constructing features, MRP projects data into the selected and extracts relations. The space of subsets of grows exponentially with the number of original and has many local optima. Therefore, the greedy search of MRP causes this method to fall in local optima when high variation exists in space of subsets. The hybrid method of Ritthoff et al applies GA for selecting to reduce the local optima problem. Feature construction is performed by arbitrarily constructing a feature from current sets of and a set of predefined constructive operators. Though features are constructed during the GA search of good the method does not apply any genetic operator or search strategy in feature construction level. Therefore, it explores the space of features without having a strategy to guide the method toward best feature. Gabret applies GA for both spaces. It consists of two GA modules: feature selection and feature construction. However, each module is performed separately and independently from the other. When high interaction exists among, attribute subset selection module cannot see the interaction among primitive. Therefore, this module often excludes some relevant F i S: Space of subsets of S i F Si : Space of functions defined over subset S i from the search space that should be used later for constructing new features. The two spaces of attribute subsets and functions over subsets are related and each one has a direct effect on the other. In order to find a good subset of it is necessary to see the relation among. An existing interaction among in subset S i is not apparent unless a function f Si, that outlines this interaction, is defined over the proper subset of. Similarly, a function can highlight the interaction among only when it is defined over the relevant. If the chosen subset of is not good enough, the best function found may not be helpful. Therefore, the two tasks, S-Search and F Si -Search, should be linked to transfer the effects of each search space into the other. 3. The Representation Language The language for representing features to show the relation among has an impact on the way the search space is conducted. The representation language should provide the capability of representing all complex functions or features of interests. This representation has an important role in convergence to optimal solution. There are two alternative groups of languages for representing features: algebraic form and non-algebraic form. By algebraic form, we mean that features are shown by means of some algebraic operators such as arithmetic or Boolean operators. Most CI methods like GCI [8], GPCI [9], Gabret [10], and the hybrid method of Ritthoff et al [11] apply this form of representation. When using this form of representation, the algebraic operators should be defined properly. GPCI uses a fix set of simple and general operators, AND and NOT, which are sufficient to represent all Boolean concepts. The use of simple operators makes the method applicable to a wide range of problems. However, a complex feature is required to capture and encapsulate the interaction using simple operators. Conversely, GCI, Gabret, and the hybrid method of Ritthoff et al apply domain-specific operators to reduce complexity; though, specifying these operators properly cannot be performed without any prior information about the target concept. In addition to the problem of defining operators, an algebraic form of representation produces an unlimited search space since any feature can be appeared in infinite forms. As an example see the following conceptually equivalent functions that are represented differently: X 1 X2 (( X1 X1) X1 1 X1) X1) X 2) 2. (((( X X

3 Therefore, a CI method needs a restriction to limit the Search space provided by this form of representation. Features can also be represented in a non-algebraic form, which means no operator is used for representing the feature. For example, for an attribute set such as {X 1, X 2 } an algebraic feature like ( X1 ( X1 can be represented by a non-algebraic feature such as <0110>, where the j th element in <0110> represents the outcome of the function for jth combination of X 1 and X 2 according to the following table: X 1 X 2 F A non-algebraic form is simpler to be applied in a CI method since there is no need to specify any operator. Besides, all features have equal degree of complexity when they are represented in a non-algebraic form. For example two algebraic features like f ( X1, = ( X1 ( X1 and f ( X 1, X 2 ) = X 1 are represented in non-algebraic form as R ( X 1, =<1001> and R ( X 1, =< 0011>, which are equivalent in terms of complexity. Since for problems with high interaction, CI needs to construct features more like the first feature in example, which are complex in algebraic form, a non-algebraic representation is preferable for these methods. Such kind of representation is applied in MRP [12], a learning method that represents features by sets of tuples using relational projection. However, MRP applies a greedy method for constructing features. As the search space for complex problems has local optima, MRP fails in some cases. 4. DCI DCI, Decomposed Constructive Induction, consists of two tasks S-Search and F Si -Search as described in Section 2. Since the space of attribute subsets S grows exponentially with the number of and has many local optima, our method applies GA for S-Search to search this space. The GA generates a population of attribute subsets and by means of genetic operators aims to converge the population to an optimal subset of. Table 1. Modified parameters of PGAPack GA Parameter S-Search Population Size 50 Max Iteration 100 Max no Change Iter. 25 Max Similarity Value 80% No. of Strings to be Replaced 46 Selection Type Proportional Crossover Type Uniform Mutation and/or Crossover And Set of Training data Figure 2. The framework of DCI GA: S-Search subset of S i F Si -Search best function and its fitness Updated data using corresponding function of the best subset For implementing GA, we use PGAPack library [13] with default parameters except those indicated in Table 1. We represent individuals (i.e., attribute subsets) by bit-strings; each bit representing presence or absence of an attribute in subset. We apply all three stopping rules of PGAPack (maximum iteration, no change in the best solution in a given number of iterations, and population too similar) as termination criteria. Recalling the need to keep the integration between two tasks, we propose to have F Si -Search inside S-Search. That is, for each subset S i in S-Search, in order to measure its fitness, the method finds the best function defined over this subset using F Si -Search (Figure 2). Then the goodness of this function determines the fitness of the subset of S i in S-Search. By this strategy we can maintain the relation between two tasks and their effects to each other, while improving each of them. F Si -Search has the main role in guiding the method towards optimal solution. Bearing in mind that the aim of F Si -Search is to construct a function over S i, the current subset of to represent interactions, we apply a non-algebraic form of representation as described in Section 3. In addition to all advantages of non-algebraic representation mentioned in Section 3, this form of representation provides the ability to extract the relations directly from data. Without losing generality, suppose that the domain is Boolean. If S i = k, DCI generates a new non-algebraic function as a string of length 2 k. The j th element of this string represents the outcome of the function for j th combination of values. The outcome is determined by the majority label in set of examples in training data that matches with the j th combination of values. If there is a gap, that is, there is not such a majority or no example in data matches with this combination of, DCI generalizes the label from all training data. For each gap a label is selected stochastically according to its probability ratio in training data. The fitness of each individual S i in S-Search depends on the goodness of the constructed feature. If the fitness of the feature F constructed by F Si -Search is Fitness Si (F), then the fitness of S i is calculated as: Si Fitness ( Si ) = FitnessSi ( F) + ε (1) S

4 where S i is the size of subset of S i and S is the size of the original set of. The first term of the formula has the major influence on fitness value of the subset S i. The last term roughly estimates the complexity of the new feature, by measuring the amount of participating in the feature. It is multiplied by ε = to have effect only when two subsets S j and S k perform equally well in F Si -Search; that is, Fitness Sj (F )= Fitness Sk (F ). Therefore, between two equally good subsets, this formula assigns a better fitness value to the smaller. The fitness of the constructed function in F Si -Search can be determined by a hypothesis-driven or a data-driven evaluation. A hypothesis-driven evaluation can be done by adding the new function to the list of and updating data, to then apply a learner like [14] and measure its accuracy. GCI, Gabret, and the hybrid method of Ritthoff et al apply this kind of evaluation. A hypothesis-driven evaluation relies on the performance of the learner. Therefore, if the learner assigns inaccurately a low or high fitness value to the feature, it will guide the algorithm to a local solution. Another possibility is to apply a data-driven evaluation formula like conditional entropy [15] to measure the amount of information needed to identify the class of a sample in data knowing the value of the new feature. A data-driven approach only depends on data, and therefore, is more reliable than a hypothesis-driven approach for guiding GA. Moreover, in GA the computation time of fitness function is very important since this function is called for every individual during each GA generation. As the computation time of a data-driven fitness function is usually less than the one for a hypothesis-driven fitness function, the former is preferable. The current version of DCI applies data-driven evaluation function. The fitness of the new feature F, for the training data T, is measured by the following formula: Fitness ( F) = Entropy( T F) (2) where Entropy(T F) is the conditional Entropy [15]. When S-Search is finished, the constructed function associated with the best individual (subset of ) in GA is added to the original set of and data are updated using the new set of. The updated data are then given to a standard learner for learning. 5. Experiments We evaluate our method by conducting experiments on artificial problems with high interactions. Similar results are expected for real-world problems where high interactions occur. We also compare DCI with other CI methods based on greedy search. Concept 5.1 Empirical Evaluation To evaluate our method, we measured the accuracy of a standard learner after constructing new features by DCI. We defined concepts over 12 Boolean a 1,, a 12 as defined in Table 2. A high interaction exists among relevant of these concepts. An important characteristic of these concepts is that there are that participate in more than one interaction, which makes the concept more difficult to be learned. Each experiment was run 10 times independently over 10 sets of data. For each experiment, we used 5% of shuffled data for training and kept the rest (95% of data) unseen for evaluating the final result. When the feature construction finished, to evaluate its performance, the accuracy of [14] on modified data after adding the constructed feature was measured using the 95% unseen data as test data. Table 3 compares DCI with two other methods. One is, a standard learner based on similarities, ran on original set of 12 (column 3). To see the need for feature construction in addition to attribute selection, we also applied on these concepts forcing it to ignore irrelevant (column 4). The second column shows number of relevant. Table 3 shows that our method considerably facilitates learning these concepts by constructing features. The Table 3. The accuracy s result average comparing to a standard learner Concept Table 2. Concepts definition No of Relevant Atts Original Atts Definition P4&P4-of-6 Parity(a 1,, a 4) and Parity(a 3,, a 6) P6&P6-of-8 Parity(a 1,, a 6) and Parity(a 3,, a 8) P3&P3&P3-of-6 Parity(a 1,, a 3) and Parity(a 3,, a 5) and Parity(a 4,, a 6) P4&P4&P4-of-6 Parity(a 1,, a 4) and Parity(a 2,, a 5) and Parity(a 3,, a 6) P4&P4&P4-of-8 Parity(a 1,, a 4) and Parity(a 3,, a 6) and Parity(a 5,, a 8) P6&P6&P6-of-8 Parity(a 1,, a 6) and Parity(a 2,, a 7) and Parity(a 3,, a 8) Relevant Atts DCI + P4&P4-of (3.6) 87.9 (5.5) 97.7 (2.2) P6&P6-of (3.3) 73.2 (2.4) 81.7 (1.3) P3&P3&P3-of (1.5) 91.7 (3.6) 99.2 (1.4) P4&P4&P4-of (0.1) 90.6 (2.5) 99.4 (0.8) P4&P4&P4-of (0.1) 87.5 (0.1) 89.2 (1.2) P6&P6&P6-of (2.2) 87.0 (1.6) 89.0 (1.1)

5 results of on relevant shows that guiding the learner to only consider interacting does not help it much to learn the concept. This implies that a preprocessing feature selection method by itself cannot help a learner when a high interaction exists among. The feature selection and feature construction should be combined together to highlight interactions. 5.2 Empirical Comparison In order to compare DCI with other CI methods, we used concepts applied in [12]. The concepts are defined over 12 Boolean a 1,, a 12 as follows: Parity-n: parity of a 1,, a n cp5-8: AND (Parity(a 5, a 6 ), Parity(a 7, a 8 )) mx6: multiplexor of a 1, a 2, a 3, a 6, a 9, a 12 mx6c-n-m: AND (a n,, a m ) if a 1 =0 and a 2 =0 OR (a n,, a m ) if a 1 =0 and a 2 =1 Parity (a n,, a m ) if a 1 =1 and a 2 =0 NOT Parity (a n,, a m ) if a 1 =1 and a 2 =1 gw5-8: the number of ones in {a 5, a 6 } is greater then sw5-8: the number of ones in {a 7, a 8 } the number of ones in {a 5, a 6 } is equal to the number of ones in {a 7, a 8 } Majn-m: among in {a n,, a m }, at least half of them are one nm-4-5: among in {a 4,, a 10 }, 4 or 5 of them are one n of m: exactly n of m are one Each experiment was ran 20 times using 5% of data for training and the rest for final evaluation and the result was evaluated by accuracy on modified data. Table 4 presents a summary of DCI s accuracy and its comparison to the other systems accuracy. The second column shows number of relevant. The last Table 4. The accuracy s result average on 5% data comparing with other CI methods Concept Relevant The best result MRP DCI + gw (0) (0) 100 (0) sw (9) (0) 100 (0) cp (7.5) (0) 100 (0) mx6c (5.8) (0) 100 (0) parity (9.3) (0) 100 (0) mj (2.8) (0.7) 99.7(1.5) mx (3.1) (7.3) 97.8 (2.1) mx6c (2.3) (3.7) 97.8 (1.8) parity (1.2) (1.6) 98.0 (1.8) mj (1.4) (5.1) 89.1 (2.7) nm (2.8) (7.6) 89.9 (1.6) 5-of (1.6) (4.7) 95.1 (1.7) 6-of (0.7) (2.3) 98.3 (1.3) parity (0.2) (1.4) 76.7 (2.7) column indicates the average accuracy after testing the result of our method by using unseen data. The results are compared with and -Rules [14], which are similarity-based learners, Fringe, Grove and Greedy3 [16], and LFC [17], which are greedy-based CI methods that apply algebraic form for representing features and MRP [12] which is a greedy CI method with a non-algebraic form of representation that uses relational projection for constructing features. Among these CI methods, Grove, Greedy3, LFC and MRP consider the decomposition of search space as explained in Section 2. In the third column of the table, we show the best results among -Rules, Fringe, Grove, Greedy3 and LFC, as reported in [12]. Numbers between parentheses indicate standard deviations. Bolds mean that with a significant level of 0.05 this accuracy is better than the others. It can be seen from Table 4 that when few are involved in interaction DCI, as most CI methods, perfectly helps the learner to learn the concept. When the number of interacting is high, the accuracy of most CI methods is scaled down. MRP slightly gives better accuracy on concepts with high number of interacting comparing to CI methods of column 4 because of its non-algebraic representation of features. The global search of DCI reduces the local optima problem. Therefore, when the number of interacting is high and the search space has high variation, the accuracy of DCI is considerably better than MRP and other methods. The concepts with small number of interacting were easy for almost all CI methods of Table 4. We reduced the number of training data to make these concepts more difficult to learn and repeated the experiments to compare the accuracies of methods on these concepts. We used 1% of data for training and kept the 99% unseen for final evaluation. This is closer to the situation of real-world problems. Table 5 shows the result of our experiments. The advantage of our method over other CI methods is now clearer. The experiments show that when a high complex interaction exists among, with a small number of training data, most CI methods with algebraic form of representation fail to construct a useful feature. MRP with non-algebraic representation can overcome this problem if there is enough training data. The small number of training data produces more variation in search space and Table 5. The accuracy s result average on 1% data comparing with other CI methods Concept Relevant The best result MRP DCI + gw (6.1) (11.1) 93.2 (7.7) sw (7.2) (18.3) 95.6 (4.1) cp (7.3) (12.1) 95.6 (4.1) mx6c (5.5) (17.4) 96.5 (3.8) parity (4) (16) 94.6 (11.1)

6 therefore, the local search of MRP fails to find optimal solution. The size of training data is less problematic for our method and therefore, DCI outperforms MRP as well as other CI methods in these experiments. Also it is important to note that in all experiments DCI successfully found the subset of interacting except one of 20 experiments over concept 6-of-7 of Table 4 and one of 20 experiments over concepts gw5-8 and parity-4 of Table Conclusion This paper has presented DCI, a new CI method based on Non-Algebraic representation, whose goal is to ease learning problems with complex attribute interaction. The design of the method decomposes the search space into two spaces of subsets of and features. The integration of two searching tasks maintains the effects of two tasks on each other while improving each of them. Other CI methods that consider these two spaces as a whole often fall in local optimum by searching over a huge space with many local optima, that is more difficult to be explored. Moreover, the genetic approach of DCI makes our method more promising than other methods in constructing adequate features when the search space is intractable and the required features are too complex to be found by greedy approaches. The non-algebraic representation of features provides a finite search space of features. The form of representation assigns an equal degree of complexity to features regardless of the size or complexity of their symbolic representation. This reduces the difficulty of constructing complex features. Besides, this form of representation provides the ability to extract features directly from training data. Empirical results proved that DCI performs well on concepts with complex interaction and outperforms other CI methods in terms of accuracy. Acknowledgement This work has been partially supported by the Spanish Interdepartmental Commission for Science Technology (CICYT), under Grants numbers TIC C02-02 and TIC References [1] L.A Rendell & R. Seshu, Learning hard concepts through constructive induction: framework and rationale, Computational Intelligence, 6, 1990, [2] A. Freitas, Understanding the crucial role of attribute interaction in data mining. Artificial Intelligence Review, 16(3), 2001, [3] N. Qian & T.J. Sejnowski, Predicting the secondary structure of globular proteins using neural network models, Journal of Molecular Biology, 202, 1988, [4] T.G. Dietterich & R.S. Michalski, Inductive learning of structural description: evaluation criteria and comparative review of selected methods, Artificial Intelligence, 16(3), 1981, [5] D.W. Aha, Incremental constructive induction: an instance-based approach, Proc. 8 th International Workshop on Machine Learning, IL, 1991, [6] J. H. Holland, Adaptation in natural and artificial systems (Ann Arbor, MI, The University of Michigan Press, 1975). [7] Z. Michalewicz, Genetic Algorithms + Data Structures = Evolution Programs (Berlin, Heidelberg, NY, Springer-Verlag, 1999). [8] H. Bensusan & I. Kuscu, Constructive Induction Using Genetic Programming, Proc. International Conference on Machine Learning, Workshop on Evolutionary Computing and Machine Learning, Bari, Italy, [9] Y. Hu, A genetic programming approach to constructive induction, Proc. 3 rd Annual Genetic Programming Conference, Madison, WI, 1998, Morgan Kauffman, [10] H. Vafaie & K. De Jong, Feature space transformation using genetic algorithms, IEEE Intelligent Systems & Their Applications, 13(2), 1998, [11] O. Ritthoff, R. Klinkenberg, S. Fischer & I. Mierswa, A hybrid approach to feature selection and generation using an evolutionary algorithm, Proc UK Workshop on Computational Intelligence. Birmingham, UK, [12] E. Pérez & L.A. Rendell, Using multidimensional projection to find relations, Proc. 12 th International Conference on Machine Learning, Tahoe City, CA, 1995, [13] D. Levine, Users guide to the PGAPack parallel genetic algorithm library, Technical Report ANL- 95/18, Argonne National Laboratory, [14] R.J. Quinlan, : Programs for machine learning, (San Mateo, CA, Morgan Kaufmann, 1993). [15] R.J. Quinlan, Induction of decision trees, Machine Learning, 1(1), 1986, [16] G. Pagallo & D. Haussler, Boolean feature discovery in empirical learning, Machine Learning, 5, 1990, [17] H. Ragavan & L.A. Rendell, Lookahead feature construction for learning hard concepts, Proc. 10 th International Conference on Machine Learning, Amherst, MA, 1993,

Binary Representations of Integers and the Performance of Selectorecombinative Genetic Algorithms

Binary Representations of Integers and the Performance of Selectorecombinative Genetic Algorithms Binary Representations of Integers and the Performance of Selectorecombinative Genetic Algorithms Franz Rothlauf Department of Information Systems University of Bayreuth, Germany franz.rothlauf@uni-bayreuth.de

More information

Constructing X-of-N Attributes with a Genetic Algorithm

Constructing X-of-N Attributes with a Genetic Algorithm Constructing X-of-N Attributes with a Genetic Algorithm Otavio Larsen 1 Alex Freitas 2 Julio C. Nievola 1 1 Postgraduate Program in Applied Computer Science 2 Computing Laboratory Pontificia Universidade

More information

CONCEPT FORMATION AND DECISION TREE INDUCTION USING THE GENETIC PROGRAMMING PARADIGM

CONCEPT FORMATION AND DECISION TREE INDUCTION USING THE GENETIC PROGRAMMING PARADIGM 1 CONCEPT FORMATION AND DECISION TREE INDUCTION USING THE GENETIC PROGRAMMING PARADIGM John R. Koza Computer Science Department Stanford University Stanford, California 94305 USA E-MAIL: Koza@Sunburn.Stanford.Edu

More information

Optimizing Feature Construction Process for Dynamic Aggregation of Relational Attributes

Optimizing Feature Construction Process for Dynamic Aggregation of Relational Attributes Journal of Computer Science 5 (11): 864-877, 2009 ISSN 1549-3636 2009 Science Publications Optimizing Feature Construction Process for Dynamic Aggregation of Relational Attributes Rayner Alfred School

More information

Evolving SQL Queries for Data Mining

Evolving SQL Queries for Data Mining Evolving SQL Queries for Data Mining Majid Salim and Xin Yao School of Computer Science, The University of Birmingham Edgbaston, Birmingham B15 2TT, UK {msc30mms,x.yao}@cs.bham.ac.uk Abstract. This paper

More information

The Genetic Algorithm for finding the maxima of single-variable functions

The Genetic Algorithm for finding the maxima of single-variable functions Research Inventy: International Journal Of Engineering And Science Vol.4, Issue 3(March 2014), PP 46-54 Issn (e): 2278-4721, Issn (p):2319-6483, www.researchinventy.com The Genetic Algorithm for finding

More information

Feature Selection Based on Relative Attribute Dependency: An Experimental Study

Feature Selection Based on Relative Attribute Dependency: An Experimental Study Feature Selection Based on Relative Attribute Dependency: An Experimental Study Jianchao Han, Ricardo Sanchez, Xiaohua Hu, T.Y. Lin Department of Computer Science, California State University Dominguez

More information

Estimating Missing Attribute Values Using Dynamically-Ordered Attribute Trees

Estimating Missing Attribute Values Using Dynamically-Ordered Attribute Trees Estimating Missing Attribute Values Using Dynamically-Ordered Attribute Trees Jing Wang Computer Science Department, The University of Iowa jing-wang-1@uiowa.edu W. Nick Street Management Sciences Department,

More information

Adaptive Crossover in Genetic Algorithms Using Statistics Mechanism

Adaptive Crossover in Genetic Algorithms Using Statistics Mechanism in Artificial Life VIII, Standish, Abbass, Bedau (eds)(mit Press) 2002. pp 182 185 1 Adaptive Crossover in Genetic Algorithms Using Statistics Mechanism Shengxiang Yang Department of Mathematics and Computer

More information

Information Fusion Dr. B. K. Panigrahi

Information Fusion Dr. B. K. Panigrahi Information Fusion By Dr. B. K. Panigrahi Asst. Professor Department of Electrical Engineering IIT Delhi, New Delhi-110016 01/12/2007 1 Introduction Classification OUTLINE K-fold cross Validation Feature

More information

Escaping Local Optima: Genetic Algorithm

Escaping Local Optima: Genetic Algorithm Artificial Intelligence Escaping Local Optima: Genetic Algorithm Dae-Won Kim School of Computer Science & Engineering Chung-Ang University We re trying to escape local optima To achieve this, we have learned

More information

A Genetic k-modes Algorithm for Clustering Categorical Data

A Genetic k-modes Algorithm for Clustering Categorical Data A Genetic k-modes Algorithm for Clustering Categorical Data Guojun Gan, Zijiang Yang, and Jianhong Wu Department of Mathematics and Statistics, York University, Toronto, Ontario, Canada M3J 1P3 {gjgan,

More information

Genetic Programming for Data Classification: Partitioning the Search Space

Genetic Programming for Data Classification: Partitioning the Search Space Genetic Programming for Data Classification: Partitioning the Search Space Jeroen Eggermont jeggermo@liacs.nl Joost N. Kok joost@liacs.nl Walter A. Kosters kosters@liacs.nl ABSTRACT When Genetic Programming

More information

Domain Independent Prediction with Evolutionary Nearest Neighbors.

Domain Independent Prediction with Evolutionary Nearest Neighbors. Research Summary Domain Independent Prediction with Evolutionary Nearest Neighbors. Introduction In January of 1848, on the American River at Coloma near Sacramento a few tiny gold nuggets were discovered.

More information

Attribute Selection with a Multiobjective Genetic Algorithm

Attribute Selection with a Multiobjective Genetic Algorithm Attribute Selection with a Multiobjective Genetic Algorithm Gisele L. Pappa, Alex A. Freitas, Celso A.A. Kaestner Pontifícia Universidade Catolica do Parana (PUCPR), Postgraduated Program in Applied Computer

More information

A Parallel Evolutionary Algorithm for Discovery of Decision Rules

A Parallel Evolutionary Algorithm for Discovery of Decision Rules A Parallel Evolutionary Algorithm for Discovery of Decision Rules Wojciech Kwedlo Faculty of Computer Science Technical University of Bia lystok Wiejska 45a, 15-351 Bia lystok, Poland wkwedlo@ii.pb.bialystok.pl

More information

Investigating the Application of Genetic Programming to Function Approximation

Investigating the Application of Genetic Programming to Function Approximation Investigating the Application of Genetic Programming to Function Approximation Jeremy E. Emch Computer Science Dept. Penn State University University Park, PA 16802 Abstract When analyzing a data set it

More information

Efficient Case Based Feature Construction

Efficient Case Based Feature Construction Efficient Case Based Feature Construction Ingo Mierswa and Michael Wurst Artificial Intelligence Unit,Department of Computer Science, University of Dortmund, Germany {mierswa, wurst}@ls8.cs.uni-dortmund.de

More information

Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic Algorithm and Particle Swarm Optimization

Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic Algorithm and Particle Swarm Optimization 2017 2 nd International Electrical Engineering Conference (IEEC 2017) May. 19 th -20 th, 2017 at IEP Centre, Karachi, Pakistan Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic

More information

Efficient SQL-Querying Method for Data Mining in Large Data Bases

Efficient SQL-Querying Method for Data Mining in Large Data Bases Efficient SQL-Querying Method for Data Mining in Large Data Bases Nguyen Hung Son Institute of Mathematics Warsaw University Banacha 2, 02095, Warsaw, Poland Abstract Data mining can be understood as a

More information

Evolution of the Discrete Cosine Transform Using Genetic Programming

Evolution of the Discrete Cosine Transform Using Genetic Programming Res. Lett. Inf. Math. Sci. (22), 3, 117-125 Available online at http://www.massey.ac.nz/~wwiims/research/letters/ Evolution of the Discrete Cosine Transform Using Genetic Programming Xiang Biao Cui and

More information

Similarity Templates or Schemata. CS 571 Evolutionary Computation

Similarity Templates or Schemata. CS 571 Evolutionary Computation Similarity Templates or Schemata CS 571 Evolutionary Computation Similarities among Strings in a Population A GA has a population of strings (solutions) that change from generation to generation. What

More information

Sequential Skewing: An Improved Skewing Algorithm

Sequential Skewing: An Improved Skewing Algorithm Sequential : An Improved Algorithm Soumya Ray sray@cs.wisc.edu Department of Computer Sciences and Department of Biostatistics and Medical Informatics, University of Wisconsin, Madison, WI 536 David Page

More information

Solving the Traveling Salesman Problem using Reinforced Ant Colony Optimization techniques

Solving the Traveling Salesman Problem using Reinforced Ant Colony Optimization techniques Solving the Traveling Salesman Problem using Reinforced Ant Colony Optimization techniques N.N.Poddar 1, D. Kaur 2 1 Electrical Engineering and Computer Science, University of Toledo, Toledo, OH, USA 2

More information

WEIGHTED K NEAREST NEIGHBOR CLASSIFICATION ON FEATURE PROJECTIONS 1

WEIGHTED K NEAREST NEIGHBOR CLASSIFICATION ON FEATURE PROJECTIONS 1 WEIGHTED K NEAREST NEIGHBOR CLASSIFICATION ON FEATURE PROJECTIONS 1 H. Altay Güvenir and Aynur Akkuş Department of Computer Engineering and Information Science Bilkent University, 06533, Ankara, Turkey

More information

Job Shop Scheduling Problem (JSSP) Genetic Algorithms Critical Block and DG distance Neighbourhood Search

Job Shop Scheduling Problem (JSSP) Genetic Algorithms Critical Block and DG distance Neighbourhood Search A JOB-SHOP SCHEDULING PROBLEM (JSSP) USING GENETIC ALGORITHM (GA) Mahanim Omar, Adam Baharum, Yahya Abu Hasan School of Mathematical Sciences, Universiti Sains Malaysia 11800 Penang, Malaysia Tel: (+)

More information

ISSN: [Keswani* et al., 7(1): January, 2018] Impact Factor: 4.116

ISSN: [Keswani* et al., 7(1): January, 2018] Impact Factor: 4.116 IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY AUTOMATIC TEST CASE GENERATION FOR PERFORMANCE ENHANCEMENT OF SOFTWARE THROUGH GENETIC ALGORITHM AND RANDOM TESTING Bright Keswani,

More information

Loopy Belief Propagation

Loopy Belief Propagation Loopy Belief Propagation Research Exam Kristin Branson September 29, 2003 Loopy Belief Propagation p.1/73 Problem Formalization Reasoning about any real-world problem requires assumptions about the structure

More information

Genetic Model Optimization for Hausdorff Distance-Based Face Localization

Genetic Model Optimization for Hausdorff Distance-Based Face Localization c In Proc. International ECCV 2002 Workshop on Biometric Authentication, Springer, Lecture Notes in Computer Science, LNCS-2359, pp. 103 111, Copenhagen, Denmark, June 2002. Genetic Model Optimization

More information

Metaheuristic Development Methodology. Fall 2009 Instructor: Dr. Masoud Yaghini

Metaheuristic Development Methodology. Fall 2009 Instructor: Dr. Masoud Yaghini Metaheuristic Development Methodology Fall 2009 Instructor: Dr. Masoud Yaghini Phases and Steps Phases and Steps Phase 1: Understanding Problem Step 1: State the Problem Step 2: Review of Existing Solution

More information

Algorithm Design (4) Metaheuristics

Algorithm Design (4) Metaheuristics Algorithm Design (4) Metaheuristics Takashi Chikayama School of Engineering The University of Tokyo Formalization of Constraint Optimization Minimize (or maximize) the objective function f(x 0,, x n )

More information

An Application of Genetic Algorithm for Auto-body Panel Die-design Case Library Based on Grid

An Application of Genetic Algorithm for Auto-body Panel Die-design Case Library Based on Grid An Application of Genetic Algorithm for Auto-body Panel Die-design Case Library Based on Grid Demin Wang 2, Hong Zhu 1, and Xin Liu 2 1 College of Computer Science and Technology, Jilin University, Changchun

More information

A New Crossover Technique for Cartesian Genetic Programming

A New Crossover Technique for Cartesian Genetic Programming A New Crossover Technique for Cartesian Genetic Programming Genetic Programming Track Janet Clegg Intelligent Systems Group, Department of Electronics University of York, Heslington York, YO DD, UK jc@ohm.york.ac.uk

More information

Design and Management of Complex Technical Processes. A Hybrid Approach to Feature Selection and. Generation Using an Evolutionary Algorithm.

Design and Management of Complex Technical Processes. A Hybrid Approach to Feature Selection and. Generation Using an Evolutionary Algorithm. REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence Methods A Hybrid Approach to Feature

More information

Inducing Parameters of a Decision Tree for Expert System Shell McESE by Genetic Algorithm

Inducing Parameters of a Decision Tree for Expert System Shell McESE by Genetic Algorithm Inducing Parameters of a Decision Tree for Expert System Shell McESE by Genetic Algorithm I. Bruha and F. Franek Dept of Computing & Software, McMaster University Hamilton, Ont., Canada, L8S4K1 Email:

More information

Simplifying OCR Neural Networks with Oracle Learning

Simplifying OCR Neural Networks with Oracle Learning SCIMA 2003 - International Workshop on Soft Computing Techniques in Instrumentation, Measurement and Related Applications Provo, Utah, USA, 17 May 2003 Simplifying OCR Neural Networks with Oracle Learning

More information

MODELLING DOCUMENT CATEGORIES BY EVOLUTIONARY LEARNING OF TEXT CENTROIDS

MODELLING DOCUMENT CATEGORIES BY EVOLUTIONARY LEARNING OF TEXT CENTROIDS MODELLING DOCUMENT CATEGORIES BY EVOLUTIONARY LEARNING OF TEXT CENTROIDS J.I. Serrano M.D. Del Castillo Instituto de Automática Industrial CSIC. Ctra. Campo Real km.0 200. La Poveda. Arganda del Rey. 28500

More information

A Case Study on Grammatical-based Representation for Regular Expression Evolution

A Case Study on Grammatical-based Representation for Regular Expression Evolution A Case Study on Grammatical-based Representation for Regular Expression Evolution Antonio González 1, David F. Barrero 2, David Camacho 1, María D. R-Moreno 2 Abstract Regular expressions, or simply regex,

More information

Fuzzy Inspired Hybrid Genetic Approach to Optimize Travelling Salesman Problem

Fuzzy Inspired Hybrid Genetic Approach to Optimize Travelling Salesman Problem Fuzzy Inspired Hybrid Genetic Approach to Optimize Travelling Salesman Problem Bindu Student, JMIT Radaur binduaahuja@gmail.com Mrs. Pinki Tanwar Asstt. Prof, CSE, JMIT Radaur pinki.tanwar@gmail.com Abstract

More information

A Memetic Heuristic for the Co-clustering Problem

A Memetic Heuristic for the Co-clustering Problem A Memetic Heuristic for the Co-clustering Problem Mohammad Khoshneshin 1, Mahtab Ghazizadeh 2, W. Nick Street 1, and Jeffrey W. Ohlmann 1 1 The University of Iowa, Iowa City IA 52242, USA {mohammad-khoshneshin,nick-street,jeffrey-ohlmann}@uiowa.edu

More information

Estimation of Distribution Algorithm Based on Mixture

Estimation of Distribution Algorithm Based on Mixture Estimation of Distribution Algorithm Based on Mixture Qingfu Zhang, Jianyong Sun, Edward Tsang, and John Ford Department of Computer Science, University of Essex CO4 3SQ, Colchester, Essex, U.K. 8th May,

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Informed Search and Exploration Chapter 4 (4.3 4.6) Searching: So Far We ve discussed how to build goal-based and utility-based agents that search to solve problems We ve also presented

More information

ET-based Test Data Generation for Multiple-path Testing

ET-based Test Data Generation for Multiple-path Testing 2016 3 rd International Conference on Engineering Technology and Application (ICETA 2016) ISBN: 978-1-60595-383-0 ET-based Test Data Generation for Multiple-path Testing Qingjie Wei* College of Computer

More information

Boolean Feature Discovery in Empirical Learning

Boolean Feature Discovery in Empirical Learning Machine Learning, 5, 71-99 (1990) 1990 Kluwer Academic Publishers. Manufactured in The Netherlands. Boolean Feature Discovery in Empirical Learning GIULIA PAGALLO (giulia@saturn.ucsc.edu) DAVID HAUSSLER

More information

A New Crossover Technique for Cartesian Genetic Programming

A New Crossover Technique for Cartesian Genetic Programming A New Crossover Technique for Cartesian Genetic Programming Genetic Programming Track Janet Clegg Intelligent Systems Group, Department of Electronics University of York, Heslington York,YODD,UK jc@ohm.york.ac.uk

More information

An Information-Theoretic Approach to the Prepruning of Classification Rules

An Information-Theoretic Approach to the Prepruning of Classification Rules An Information-Theoretic Approach to the Prepruning of Classification Rules Max Bramer University of Portsmouth, Portsmouth, UK Abstract: Keywords: The automatic induction of classification rules from

More information

Parameter Control of Genetic Algorithms by Learning and Simulation of Bayesian Networks

Parameter Control of Genetic Algorithms by Learning and Simulation of Bayesian Networks Submitted Soft Computing Parameter Control of Genetic Algorithms by Learning and Simulation of Bayesian Networks C. Bielza,*, J.A. Fernández del Pozo, P. Larrañaga Universidad Politécnica de Madrid, Departamento

More information

Informed search algorithms. (Based on slides by Oren Etzioni, Stuart Russell)

Informed search algorithms. (Based on slides by Oren Etzioni, Stuart Russell) Informed search algorithms (Based on slides by Oren Etzioni, Stuart Russell) The problem # Unique board configurations in search space 8-puzzle 9! = 362880 15-puzzle 16! = 20922789888000 10 13 24-puzzle

More information

An evolutionary annealing-simplex algorithm for global optimisation of water resource systems

An evolutionary annealing-simplex algorithm for global optimisation of water resource systems FIFTH INTERNATIONAL CONFERENCE ON HYDROINFORMATICS 1-5 July 2002, Cardiff, UK C05 - Evolutionary algorithms in hydroinformatics An evolutionary annealing-simplex algorithm for global optimisation of water

More information

Improving Tree-Based Classification Rules Using a Particle Swarm Optimization

Improving Tree-Based Classification Rules Using a Particle Swarm Optimization Improving Tree-Based Classification Rules Using a Particle Swarm Optimization Chi-Hyuck Jun *, Yun-Ju Cho, and Hyeseon Lee Department of Industrial and Management Engineering Pohang University of Science

More information

arxiv: v1 [cs.ne] 22 Mar 2016

arxiv: v1 [cs.ne] 22 Mar 2016 Adaptive Parameter Selection in Evolutionary Algorithms by Reinforcement Learning with Dynamic Discretization of Parameter Range arxiv:1603.06788v1 [cs.ne] 22 Mar 2016 ABSTRACT Arkady Rost ITMO University

More information

REAL-CODED GENETIC ALGORITHMS CONSTRAINED OPTIMIZATION. Nedim TUTKUN

REAL-CODED GENETIC ALGORITHMS CONSTRAINED OPTIMIZATION. Nedim TUTKUN REAL-CODED GENETIC ALGORITHMS CONSTRAINED OPTIMIZATION Nedim TUTKUN nedimtutkun@gmail.com Outlines Unconstrained Optimization Ackley s Function GA Approach for Ackley s Function Nonlinear Programming Penalty

More information

WRAPPER feature selection method with SIPINA and R (RWeka package). Comparison with a FILTER approach implemented into TANAGRA.

WRAPPER feature selection method with SIPINA and R (RWeka package). Comparison with a FILTER approach implemented into TANAGRA. 1 Topic WRAPPER feature selection method with SIPINA and R (RWeka package). Comparison with a FILTER approach implemented into TANAGRA. Feature selection. The feature selection 1 is a crucial aspect of

More information

PASS EVALUATING IN SIMULATED SOCCER DOMAIN USING ANT-MINER ALGORITHM

PASS EVALUATING IN SIMULATED SOCCER DOMAIN USING ANT-MINER ALGORITHM PASS EVALUATING IN SIMULATED SOCCER DOMAIN USING ANT-MINER ALGORITHM Mohammad Ali Darvish Darab Qazvin Azad University Mechatronics Research Laboratory, Qazvin Azad University, Qazvin, Iran ali@armanteam.org

More information

Genetic Programming. and its use for learning Concepts in Description Logics

Genetic Programming. and its use for learning Concepts in Description Logics Concepts in Description Artificial Intelligence Institute Computer Science Department Dresden Technical University May 29, 2006 Outline Outline: brief introduction to explanation of the workings of a algorithm

More information

Thesis Overview KNOWLEDGE EXTRACTION IN LARGE DATABASES USING ADAPTIVE STRATEGIES

Thesis Overview KNOWLEDGE EXTRACTION IN LARGE DATABASES USING ADAPTIVE STRATEGIES Thesis Overview KNOWLEDGE EXTRACTION IN LARGE DATABASES USING ADAPTIVE STRATEGIES Waldo Hasperué Advisor: Laura Lanzarini Facultad de Informática, Universidad Nacional de La Plata PhD Thesis, March 2012

More information

Evolutionary Methods for State-based Testing

Evolutionary Methods for State-based Testing Evolutionary Methods for State-based Testing PhD Student Raluca Lefticaru Supervised by Florentin Ipate University of Piteşti, Romania Department of Computer Science Outline Motivation Search-based software

More information

Metaheuristic Optimization with Evolver, Genocop and OptQuest

Metaheuristic Optimization with Evolver, Genocop and OptQuest Metaheuristic Optimization with Evolver, Genocop and OptQuest MANUEL LAGUNA Graduate School of Business Administration University of Colorado, Boulder, CO 80309-0419 Manuel.Laguna@Colorado.EDU Last revision:

More information

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra Pattern Recall Analysis of the Hopfield Neural Network with a Genetic Algorithm Susmita Mohapatra Department of Computer Science, Utkal University, India Abstract: This paper is focused on the implementation

More information

Genetically Enhanced Parametric Design for Performance Optimization

Genetically Enhanced Parametric Design for Performance Optimization Genetically Enhanced Parametric Design for Performance Optimization Peter VON BUELOW Associate Professor, Dr. -Ing University of Michigan Ann Arbor, USA pvbuelow@umich.edu Peter von Buelow received a BArch

More information

Analyzing Outlier Detection Techniques with Hybrid Method

Analyzing Outlier Detection Techniques with Hybrid Method Analyzing Outlier Detection Techniques with Hybrid Method Shruti Aggarwal Assistant Professor Department of Computer Science and Engineering Sri Guru Granth Sahib World University. (SGGSWU) Fatehgarh Sahib,

More information

Using Genetic Algorithms to optimize ACS-TSP

Using Genetic Algorithms to optimize ACS-TSP Using Genetic Algorithms to optimize ACS-TSP Marcin L. Pilat and Tony White School of Computer Science, Carleton University, 1125 Colonel By Drive, Ottawa, ON, K1S 5B6, Canada {mpilat,arpwhite}@scs.carleton.ca

More information

A Two-level Learning Method for Generalized Multi-instance Problems

A Two-level Learning Method for Generalized Multi-instance Problems A wo-level Learning Method for Generalized Multi-instance Problems Nils Weidmann 1,2, Eibe Frank 2, and Bernhard Pfahringer 2 1 Department of Computer Science University of Freiburg Freiburg, Germany weidmann@informatik.uni-freiburg.de

More information

MAXIMUM LIKELIHOOD ESTIMATION USING ACCELERATED GENETIC ALGORITHMS

MAXIMUM LIKELIHOOD ESTIMATION USING ACCELERATED GENETIC ALGORITHMS In: Journal of Applied Statistical Science Volume 18, Number 3, pp. 1 7 ISSN: 1067-5817 c 2011 Nova Science Publishers, Inc. MAXIMUM LIKELIHOOD ESTIMATION USING ACCELERATED GENETIC ALGORITHMS Füsun Akman

More information

Using Decision Boundary to Analyze Classifiers

Using Decision Boundary to Analyze Classifiers Using Decision Boundary to Analyze Classifiers Zhiyong Yan Congfu Xu College of Computer Science, Zhejiang University, Hangzhou, China yanzhiyong@zju.edu.cn Abstract In this paper we propose to use decision

More information

Informed search algorithms. Chapter 4

Informed search algorithms. Chapter 4 Informed search algorithms Chapter 4 Material Chapter 4 Section 1 - Exclude memory-bounded heuristic search 3 Outline Best-first search Greedy best-first search A * search Heuristics Local search algorithms

More information

Research on Applications of Data Mining in Electronic Commerce. Xiuping YANG 1, a

Research on Applications of Data Mining in Electronic Commerce. Xiuping YANG 1, a International Conference on Education Technology, Management and Humanities Science (ETMHS 2015) Research on Applications of Data Mining in Electronic Commerce Xiuping YANG 1, a 1 Computer Science Department,

More information

Research Article Path Planning Using a Hybrid Evolutionary Algorithm Based on Tree Structure Encoding

Research Article Path Planning Using a Hybrid Evolutionary Algorithm Based on Tree Structure Encoding e Scientific World Journal, Article ID 746260, 8 pages http://dx.doi.org/10.1155/2014/746260 Research Article Path Planning Using a Hybrid Evolutionary Algorithm Based on Tree Structure Encoding Ming-Yi

More information

Fuzzy Partitioning with FID3.1

Fuzzy Partitioning with FID3.1 Fuzzy Partitioning with FID3.1 Cezary Z. Janikow Dept. of Mathematics and Computer Science University of Missouri St. Louis St. Louis, Missouri 63121 janikow@umsl.edu Maciej Fajfer Institute of Computing

More information

Deriving Trading Rules Using Gene Expression Programming

Deriving Trading Rules Using Gene Expression Programming 22 Informatica Economică vol. 15, no. 1/2011 Deriving Trading Rules Using Gene Expression Programming Adrian VISOIU Academy of Economic Studies Bucharest - Romania Economic Informatics Department - collaborator

More information

HAM: A HYBRID ALGORITHM PREDICTION BASED MODEL FOR ELECTRICITY DEMAND

HAM: A HYBRID ALGORITHM PREDICTION BASED MODEL FOR ELECTRICITY DEMAND HAM: A HYBRID ALGORITHM PREDICTION BASED MODEL FOR ELECTRICITY DEMAND 1 WAHAB MUSA, 2 SALMAWATY TANSA 1 Assoc. Prof., Department of Electrical Engineering, UNIVERSITAS NEGERI GORONTALO 2 Lektor., Department

More information

Applied Cloning Techniques for a Genetic Algorithm Used in Evolvable Hardware Design

Applied Cloning Techniques for a Genetic Algorithm Used in Evolvable Hardware Design Applied Cloning Techniques for a Genetic Algorithm Used in Evolvable Hardware Design Viet C. Trinh vtrinh@isl.ucf.edu Gregory A. Holifield greg.holifield@us.army.mil School of Electrical Engineering and

More information

e-ccc-biclustering: Related work on biclustering algorithms for time series gene expression data

e-ccc-biclustering: Related work on biclustering algorithms for time series gene expression data : Related work on biclustering algorithms for time series gene expression data Sara C. Madeira 1,2,3, Arlindo L. Oliveira 1,2 1 Knowledge Discovery and Bioinformatics (KDBIO) group, INESC-ID, Lisbon, Portugal

More information

IJMIE Volume 2, Issue 9 ISSN:

IJMIE Volume 2, Issue 9 ISSN: Dimensionality Using Optimization Algorithm for High Dimensional Data Clustering Saranya.S* Dr.Punithavalli.M** Abstract: This paper present an efficient approach to a feature selection problem based on

More information

Random Search Report An objective look at random search performance for 4 problem sets

Random Search Report An objective look at random search performance for 4 problem sets Random Search Report An objective look at random search performance for 4 problem sets Dudon Wai Georgia Institute of Technology CS 7641: Machine Learning Atlanta, GA dwai3@gatech.edu Abstract: This report

More information

Heuristic Optimisation

Heuristic Optimisation Heuristic Optimisation Revision Lecture Sándor Zoltán Németh http://web.mat.bham.ac.uk/s.z.nemeth s.nemeth@bham.ac.uk University of Birmingham S Z Németh (s.nemeth@bham.ac.uk) Heuristic Optimisation University

More information

Using Text Learning to help Web browsing

Using Text Learning to help Web browsing Using Text Learning to help Web browsing Dunja Mladenić J.Stefan Institute, Ljubljana, Slovenia Carnegie Mellon University, Pittsburgh, PA, USA Dunja.Mladenic@{ijs.si, cs.cmu.edu} Abstract Web browsing

More information

CS 331: Artificial Intelligence Local Search 1. Tough real-world problems

CS 331: Artificial Intelligence Local Search 1. Tough real-world problems CS 331: Artificial Intelligence Local Search 1 1 Tough real-world problems Suppose you had to solve VLSI layout problems (minimize distance between components, unused space, etc.) Or schedule airlines

More information

A Genetic Algorithm for Minimum Tetrahedralization of a Convex Polyhedron

A Genetic Algorithm for Minimum Tetrahedralization of a Convex Polyhedron A Genetic Algorithm for Minimum Tetrahedralization of a Convex Polyhedron Kiat-Choong Chen Ian Hsieh Cao An Wang Abstract A minimum tetrahedralization of a convex polyhedron is a partition of the convex

More information

Learning Adaptive Parameters with Restricted Genetic Optimization Method

Learning Adaptive Parameters with Restricted Genetic Optimization Method Learning Adaptive Parameters with Restricted Genetic Optimization Method Santiago Garrido and Luis Moreno Universidad Carlos III de Madrid, Leganés 28911, Madrid (Spain) Abstract. Mechanisms for adapting

More information

Lecture 6: The Building Block Hypothesis. Genetic Algorithms and Genetic Programming Lecture 6. The Schema Theorem Reminder

Lecture 6: The Building Block Hypothesis. Genetic Algorithms and Genetic Programming Lecture 6. The Schema Theorem Reminder Lecture 6: The Building Block Hypothesis 1 Genetic Algorithms and Genetic Programming Lecture 6 Gillian Hayes 9th October 2007 The Building Block Hypothesis Experimental evidence for the BBH The Royal

More information

Dynamic Clustering of Data with Modified K-Means Algorithm

Dynamic Clustering of Data with Modified K-Means Algorithm 2012 International Conference on Information and Computer Networks (ICICN 2012) IPCSIT vol. 27 (2012) (2012) IACSIT Press, Singapore Dynamic Clustering of Data with Modified K-Means Algorithm Ahamed Shafeeq

More information

IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 5, NO. 1, FEBRUARY

IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 5, NO. 1, FEBRUARY IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 5, NO. 1, FEBRUARY 2001 41 Brief Papers An Orthogonal Genetic Algorithm with Quantization for Global Numerical Optimization Yiu-Wing Leung, Senior Member,

More information

Genetic Algorithms and Genetic Programming Lecture 7

Genetic Algorithms and Genetic Programming Lecture 7 Genetic Algorithms and Genetic Programming Lecture 7 Gillian Hayes 13th October 2006 Lecture 7: The Building Block Hypothesis The Building Block Hypothesis Experimental evidence for the BBH The Royal Road

More information

Time Complexity Analysis of the Genetic Algorithm Clustering Method

Time Complexity Analysis of the Genetic Algorithm Clustering Method Time Complexity Analysis of the Genetic Algorithm Clustering Method Z. M. NOPIAH, M. I. KHAIRIR, S. ABDULLAH, M. N. BAHARIN, and A. ARIFIN Department of Mechanical and Materials Engineering Universiti

More information

Stochastic propositionalization of relational data using aggregates

Stochastic propositionalization of relational data using aggregates Stochastic propositionalization of relational data using aggregates Valentin Gjorgjioski and Sašo Dzeroski Jožef Stefan Institute Abstract. The fact that data is already stored in relational databases

More information

GA is the most popular population based heuristic algorithm since it was developed by Holland in 1975 [1]. This algorithm runs faster and requires les

GA is the most popular population based heuristic algorithm since it was developed by Holland in 1975 [1]. This algorithm runs faster and requires les Chaotic Crossover Operator on Genetic Algorithm Hüseyin Demirci Computer Engineering, Sakarya University, Sakarya, 54187, Turkey Ahmet Turan Özcerit Computer Engineering, Sakarya University, Sakarya, 54187,

More information

The Ohio State University Columbus, Ohio, USA Universidad Autónoma de Nuevo León San Nicolás de los Garza, Nuevo León, México, 66450

The Ohio State University Columbus, Ohio, USA Universidad Autónoma de Nuevo León San Nicolás de los Garza, Nuevo León, México, 66450 Optimization and Analysis of Variability in High Precision Injection Molding Carlos E. Castro 1, Blaine Lilly 1, José M. Castro 1, and Mauricio Cabrera Ríos 2 1 Department of Industrial, Welding & Systems

More information

Computational Intelligence

Computational Intelligence Computational Intelligence Module 6 Evolutionary Computation Ajith Abraham Ph.D. Q What is the most powerful problem solver in the Universe? ΑThe (human) brain that created the wheel, New York, wars and

More information

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM 20 CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM 2.1 CLASSIFICATION OF CONVENTIONAL TECHNIQUES Classical optimization methods can be classified into two distinct groups:

More information

CLOSED LOOP SYSTEM IDENTIFICATION USING GENETIC ALGORITHM

CLOSED LOOP SYSTEM IDENTIFICATION USING GENETIC ALGORITHM CLOSED LOOP SYSTEM IDENTIFICATION USING GENETIC ALGORITHM Lucchesi Alejandro (a), Campomar Guillermo (b), Zanini Aníbal (c) (a,b) Facultad Regional San Nicolás Universidad Tecnológica Nacional (FRSN-UTN),

More information

The Simple Genetic Algorithm Performance: A Comparative Study on the Operators Combination

The Simple Genetic Algorithm Performance: A Comparative Study on the Operators Combination INFOCOMP 20 : The First International Conference on Advanced Communications and Computation The Simple Genetic Algorithm Performance: A Comparative Study on the Operators Combination Delmar Broglio Carvalho,

More information

Mining High Order Decision Rules

Mining High Order Decision Rules Mining High Order Decision Rules Y.Y. Yao Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 e-mail: yyao@cs.uregina.ca Abstract. We introduce the notion of high

More information

Using Genetic Algorithm with Triple Crossover to Solve Travelling Salesman Problem

Using Genetic Algorithm with Triple Crossover to Solve Travelling Salesman Problem Proc. 1 st International Conference on Machine Learning and Data Engineering (icmlde2017) 20-22 Nov 2017, Sydney, Australia ISBN: 978-0-6480147-3-7 Using Genetic Algorithm with Triple Crossover to Solve

More information

A HYBRID APPROACH IN GENETIC ALGORITHM: COEVOLUTION OF THREE VECTOR SOLUTION ENCODING. A CASE-STUDY

A HYBRID APPROACH IN GENETIC ALGORITHM: COEVOLUTION OF THREE VECTOR SOLUTION ENCODING. A CASE-STUDY A HYBRID APPROACH IN GENETIC ALGORITHM: COEVOLUTION OF THREE VECTOR SOLUTION ENCODING. A CASE-STUDY Dmitriy BORODIN, Victor GORELIK, Wim DE BRUYN and Bert VAN VRECKEM University College Ghent, Ghent, Belgium

More information

Challenges and Interesting Research Directions in Associative Classification

Challenges and Interesting Research Directions in Associative Classification Challenges and Interesting Research Directions in Associative Classification Fadi Thabtah Department of Management Information Systems Philadelphia University Amman, Jordan Email: FFayez@philadelphia.edu.jo

More information

Modified K-Means Algorithm for Genetic Clustering

Modified K-Means Algorithm for Genetic Clustering 24 Modified K-Means Algorithm for Genetic Clustering Mohammad Babrdel Bonab Islamic Azad University Bonab Branch, Iran Summary The K-Means Clustering Approach is one of main algorithms in the literature

More information

Learning Fuzzy Rules by Evolution for Mobile Agent Control

Learning Fuzzy Rules by Evolution for Mobile Agent Control Learning Fuzzy Rules by Evolution for Mobile Agent Control George Chronis, James Keller and Marjorie Skubic Dept. of Computer Engineering and Computer Science, University of Missouri-Columbia, email: gchronis@ece.missouri.edu

More information

EVOLVING LEGO. Exploring the impact of alternative encodings on the performance of evolutionary algorithms. 1. Introduction

EVOLVING LEGO. Exploring the impact of alternative encodings on the performance of evolutionary algorithms. 1. Introduction N. Gu, S. Watanabe, H. Erhan, M. Hank Haeusler, W. Huang, R. Sosa (eds.), Rethinking Comprehensive Design: Speculative Counterculture, Proceedings of the 19th International Conference on Computer- Aided

More information