INFORMATION DRIVEN OPTIMIZATION SEARCH FILTER: PREDICTING TABU REGIONS. Matthew H. Jones

Size: px
Start display at page:

Download "INFORMATION DRIVEN OPTIMIZATION SEARCH FILTER: PREDICTING TABU REGIONS. Matthew H. Jones"

Transcription

1 Proceedings of the Systems and Information Engineering Design Symposium Matthew H. Jones, Stephen D. Patek, and Barbara E. Tawney, eds. INFORMATION DRIVEN OPTIMIZATION SEARCH FILTER: PREDICTING TABU REGIONS Matthew H. Jones Department of Systems and Information Engineering University of Virginia ABSTRACT Many search techniques fail to account for the information obtained from previous objective function evaluations when determining a new set of control parameters. This paper presents an empirical study of a neural net prescreener using a random and grid search however one may use the pre-screener in combination with any search procedure. A single neural network model is exted through the usage hierarchical clustering to organize the search space into groups with a corresponding neural network model. Empirical tests indicate that a neural network prescreener is beneficial in significantly reducing the number of probes with a minimal cost to accuracy, an acceptable trade-off given the high cost of executing complex objective functions. Specifically, a single neural network model is optimal given a random search. Hierarchical clustering using m separate neural network models outperforms, in terms of deviance and number of probes, the usage of a single model with respect to a grid search. The grid search provides a broader coverage of the search space, yielding more information regarding the search space. INTRODUCTION Many techniques fail to account for the information obtained from previous objective function evaluations when determining a new set of control parameters. Searches may blindly search the feasible control space, exhibit a level one Markovian process, or keep a record of previously explored regions when selecting the new vector of control parameters for future search. Historically the determination of the next search area has been a computational simple task however this paradigm is slowly shifting. Complex computer simulations models execute in times magnitudes larger than the time to determine the next control parameters to search. Time spent in intelligently selecting the next search region, to minimize the number of objective function evaluations, drastically reduces the overall time for optimization. Information discovery and data mining principles provide insight into the global behavior of the objective function response surface. Each new objective function evaluation yields further insight into the response surface, identifying areas of potential optimum, considering the information obtained from all previous objective function evaluations. OptQuest, a commercially available optimization package, performs just such a task using a neural network pre-screener to identifying potentially inferior sets of control points prior to exping the resources to evaluate the objective function [, ]. OptQuest s neural net pre-screener is proprietary knowledge. The paper presents an empirical study of a neural net pre-screener using a random and grid search however one may use the pre-screener in combination with any search procedure. A single neural net model is fit over the entire search space. A set of points must be calculated to initially populate the neural net as well as the metric for disqualifying a control point based on the neural net model, also called the risk level. Both quantities require sensitivity analysis to gain insight and in making general recommations regarding usage of the pre-screener over a general class of problems. The mean squared error and mean absolute deviation error metrics act as the fit minimization criterion in the neural net prediction model. In all instances, the number of probes and deviation from the best answer with and without the pre-screener act as metrics for comparison among implementations. Lastly, a single neural net model may not provide the best prediction over the entire search space. A hierarchical clustering procedure organizes the clusters into groups and a separate neural net model fit over each set of clusters. A potential control vector is classified into one of the m clusters and a prediction generated from the specific neural net model. A preliminary study is presented. SELECTION OF NEURAL NETWORKS Neural networks contain simple elements or neurons operating in parallel inspired by the biological nervous system. The structure and prediction generated from a neural net is determined by the connections between elements. The weights and connections of the neurons adapt to match a supervised learning pattern between the input and output. Feed forward backpropogation neural nets were selected given their ability to provide reasonable predictions when

2 presented with non-training data. For more information regarding neural networks and the backpropogation method please see [, ]. MATLAB s built in neural network toolbox and hierarchical clustering functions provide an excellent tool for coding the problem. All networks have two layers with the first layer having ten nodes with a log-sigmoid transfer function and the second layer containing one node with a linear transfer function. The performance function is either mean squared error (MSE, an L error metric) or mean absolute deviation (MAD, an L error metric). A single outlier using the MSE performance measure is heavily penalized and highly influences the other points, a potentially undesirable property. The best minimum is defined as the best minimum value observed in the search. The predicted output is the value predicted, for the potential next set of control parameters using the neural net model. This value is then normalized by the best minimum. If this value is less than some threshold, we accept and if not the point is thrown out and not executed over the loss function.. Random Search.. Determination of Risk Level FITTING A SINGLE NEURAL NETWORK OVER THE SEARH SPACE The following function is used in fitting a single neural network model over the entire search space. Empirical analysis over this problem yields insight into determining an appropriate number of points to initially the fit the network, determine the risk level, and the performance criterion (MSE or MAD). The function is as follows: [( x x + x ) + ( x x x )] L ( x, x) = + () T we wish to minimize L ( x, x ); Θ [, ] with * θ = [. 9,. 9]. The function is multi-modal with four local minimum, one of which is the global optimum. It may be seen below: Local Minima Local Minima Local Minima Deviation Fliter/No-Filter Changing Risk Level - Random Threshold Value 9 Figure. Risk Level Random Search on MAE Average MSE Average MAE SD MSE SD The above figure summarizes the threshold value as a function of the deviation obtained by implementing the neural net pre-filter. Each point represents the performance metric over replications. The solid lines represent the average over the replications for the MSE and MAD performance criterion. The dashed lines denote the standard deviation over the replications for the MSE and MAD performance criterion. As the threshold value increases, the more likely a control point is to be accepted and thus the greater number of probes into the response surface. It is interesting to note that the mean and standard deviation are similar. In this empirical case, neither the MSE or MAD performance criterion dominate over all values of test threshold values. Global Minima Figure. Contour Plot of Response Surface The risk level is defined as follows: Best_Min. Pred_Output / Best_Min. < τ ()

3 Changing Risk Level.. Determination of Initial Sample Size. Number of Probes (Possible ) MAE Average MSE Average MAE SD MSE SD Number of Probes (Possible ) Threshold Figure. Risk Level Random Search on Number Probes The above figure displays the threshold value as a function of the number of probes. As the threshold value increases, the number of probes increases. The standard deviation over the samples is displayed as the two dashed lines. These results mirror the deviance chart displayed in Figure. As the threshold value increases, the deviance goes down however this requires more surface probes. The slope of the averages for MSE and MAD indicate a large performance jump from moving from a threshold value of. to.. The deviance seen in Figure from. to. is similar for a % reduction in probes. A threshold of. in this instance appears optimal in reducing the number of probes while minimizing the deviance. The plot below plots the mean deviance and number of probes for the MSE and MAD case. In this case, MSE and MAD exhibit similar performance measures. Intial Sample Size MAE Number Probes MSE Number Probes MAE MSE Figure. Initial Sample Size Random Search Summary The above plot displays the sensitivity analysis over the number of probes, the dashed lines representing the deviance, solid lines the number of probes. Decreasing the initial sample size decreases the number of probes however increases the deviance and vice versa. In general, the MAE error criterion exhibits minimal deviance with a fewer number of probes than does the MSE criterion across the initial sample size. Given this empirical results and the desire to not heavily penalize outliers, MAD is the preferred performance metric for minimization when fitting a neural network model.. Grid Search 9.9 Number of Probes % Reduction in Possible Number of Probes Threshold Value MAE MSE MAE Number of Probes MSE Number of Probes Figure. Risk Level Random Search Summary. Grid Size (NxN) % Reduction in Possible Number of Probes Figure. Initial Sample Size Random Search Summary The above plot, taken over a series of grid searches with varying grid sizes, displays the deviance and probe reduction as a function of the grid size. The deviance from having the neural net pre-screener and not have the prescreener decreases as the grid size increases. The percent reduction in probes increases as the grid size increases. The

4 number of probes increases however the percentage of probes over the number of possible probes decreases. It is best to maximize the available probe budget in a grid search to minimize the deviance. FITTING MULTIPLE NEURAL NETWORKS OVER THE SEARCH SPACE A single network may not yield an accurate prediction over the entire response surface, particularly if the surface is multi-modal and large jumps in the response occur with a small change in the control parameters. This occurs with a steep gradient, may indicate a local optimum, and is difficult to predict. The chart below shows a second order polynomial fit over a cross section of the test function. A single second order fit yields a poor fit given the multimodal nature of the data. This provides an incentive to investigate the impact of fitting multiple neural net models over the response surface. The difficulty comes in determining how to separate regions into partitions and to classify a new control parameter into one of m possible partitions for prediction. Hierarchical clustering provides a solution.. AVELINK Clustering using Euclidean Distances with Clusters Average link clustering is implored with the Euclidean distance metric. In this empirical test, either three or four separate clusters are chosen. Figure 9 displays a sample drogram and displays at what distances individual points or clusters join to form larger clusters. The number of intersections, with the vertical lines, formed by drawing a horizontal line across the drogram determines the number of clusters formed. Thus in Figure 9 the placement of the horizontal line across the figure indicates the formation of four clusters Distance at Which Clusters/Points Join - Figure. Fitting One Second Order Polynomial Model Cluster Number Figure 9. Sample Drogram Fitting two separate second order polynomial model exactly matches data, yielding far better predictive power... Random Search Figure. Fitting Two Second Order Polynomial Models In the empirical test for a random search, points are probed to form the initial clusters. A neural net model is fit over each of either three of four groups of clusters. The generation of a new point by the random search requires the following steps:. Generate new point (randomly or next in algorithm) for potential objective function evaluation. Calculate the centroid of each cluster density by average the x and x s for each of the m clusters. Calculate the Euclidean distance between each of the m cluster centroids and the new point. Make the prediction based on the points assignment to the neural networks associated with the selected cluster. Accept the new point or reject based on the associated risk level. If the new point is accept, re-calculate the m clusters and update the m neural networks.

5 The plot below displays a sample set of clusters with the larger shadowed points denoting the centroid Figure. Example Clusters for Grid Search -. Number of Probes - Figure. Example Clusters for Random Search Clusters Clusters Linear ( Clusters) Linear ( Clusters) Figure. Random Search Results The above plot shows the deviance as a function of the number of probes in the case of using three of four clusters. A least squares regression line for the cluster cases suggests deviance and the number of probes are not related and in the cluster case, a decreases in the number of probes denotes an increase in the deviance. Comparisons with usage of a single neural net model occurs in Section. Grid Size Probes Used / Possible Number Probes Figure. and Probes Used as a Function of Grid Size with Clusters The above plot displays the deviance as a function of grid size and the number of probes used over the possible number of probes. The percentage of possible probes used decreases as the grid size decreases and the deviance decreases, similar results to that of imploring a single neural network model. Comparisons with using a single neural net model occurs in Section Probes Used / Possible Number Probes.. Grid Search Four separate clusters are fit over the search region imploring a grid search. The chart below displays the clusters with the respective centroid.

6 COMPARISONS. Random Search superior for a grid search compared to a single neural network model. CONCLUSIONS from Best Optimum in Probes 9 - Number of Probes Rand Rand Rand Rand Neural Networks NN with Clustering Figure. Random Search Comparison The above plot displays empirical trials of a random search with and without the neural network pre-screener as well as neural network model imploring hierarchical clustering. Fitting a neural model outperforms half of the random runs not using a neural network pre-screener and performs similarly to the other two runs. In general, NN with clustering is outperformed by the a single neural network model based on the number of probes and accuracy. The empirical tests indicate that a neural network prescreener is beneficial in significantly reducing the number of probes with a minimal cost to accuracy, an acceptable trade-off given the high cost of executing complex objective functions. Specifically, a single neural network model, in this empirical test, is appropriate given a random search. The grid search provides a broader coverage of the search space, yielding more information regarding the search space. The number of points within each of the m clusters is more uniformly distributed than the clusters created in a random search. Hierarchical clustering using m separate neural network models outperforms, in terms of deviance and number of probes, the usage of a single model when the underlying search process is a grid search. This provides empirical evidence that information driven search procedures may provide significant performance enhancements in optimization search procedures. Clearly more research is required in the broader applicability of the tool however the widespread usage and acceptability of OptQuest, imploring a neural network pre-filter, provides promising results. REFERENCES. Grid Search 9 9 Grid Size - NN - NN with Clustering Probe Reduction - NN Probe Reduction - NN with Clustering Figure. Grid Search Comparison % Possible Probes [] M. Laguna, "Metaheuristic Optimization with Evolver, Genocop and OptQuest," l, 99. [] F. K. Glover, James P. ; Laguna, Manuel, "The OptQuest Approach to Crystal Ball Simulation Optimization," l. [] R. E. Uhrig, "Introduction to artificial neural networks," presented at Industrial Electronics, Control, and Instrumentation, 99., Proceedings of the 99 IEEE IECON st International Conference on, 99. [] A. K. Jain, J. Mao, and K. M. Mohiuddin, "Artificial neural networks: a tutorial," Computer, vol. 9, pp. -, 99. The above figure summarizes the grid search with and without the usage of hierarchical clustering. The lines representing the deviance as a function of the grid size nearly coincide between the two methods. The lines representing the probe reduction denote that neural networks with hierarchical clustering require far fewer probes with the same deviance level. This empirically shows that creating m clusters and fitting m neural networks over each cluster is AUTHOR BIOGRAPHIES MATTHEW H. JONES is a Ph.D student in the Systems and Information Engineering Department under the direction of Professor William T. Scherer. He received his B.S and M.S. from the University of Virginia in Systems Engineering. He begins employment with The Aerospace Cor-

7 poration in Chantilly, VA in the Systems Architecture and Engineering Department at the of the summer session to remotely write his dissertation. He currently serves as the head teaching assistant for the Accelerated Masters Program in Systems Engineering. He is a member of INFORMS and INCOSE. APPENDIX A: MATLAB CODE SNIPPETS % Define a Neural Network net = newff([-, ; -, ], [ ], {'logsig','purelin'}); net.performfcn = 'mae'; % Perform Initial Clustering Y = pdist(predictor'); Z = linkage(y, 'average'); T = cluster(z, ); for i = :length(predictor') if (T(i) == ) X = size(cluster); Cluster(X(,) +,:) = Predictor_T(i, :); Cluster_R(X(,)+) = Response(i); if (T(i) == ) X = size(cluster); Cluster(X(,) +,:) = Predictor_T(i, :); Cluster_R(X(,)+) = Response(i); if (T(i) == ) X = size(cluster); Cluster(X(,) +,:) = Predictor_T(i, :); Cluster_R(X(,)+) = Response(i); if (T(i) == ) X = size(cluster); Cluster(X(,) +,:) = Predictor_T(i, :); Cluster_R(X(,)+) = Response(i); % Compute the centroids A = []; A(max(T),) = ; N = []; N(max(T),) = ; Centroids = []; Centroids(max(T), ) = ; for i = :length(t) N(T(i)) = N(T(i)) + ; for j=: A(T(i),j) = A(T(i),j) + Predictor_T(i,j); for i = :max(t) for j=: Centroids(i,j) = A(i,j) / N(i); % Determine the cluster closet to the current centroids min = ; min_cluster = ; for i = :max(t) Distance = []; Distance(,) = x; Distance(,) = x; Distance(,) = Centroids(i, ); Distance(,) = Centroids(i, ); distance = pdist(distance, DISTANCE_METRIC); if (distance < min) min_cluster = i; min = distance; % Train the neural nets over each cluster if (update_net == ) if (length(cluster) > ) net = train(net, Cluster', Cluster_R); if (length(cluster) > ) net = train(net, Cluster', Cluster_R); if (length(cluster) > ) net = train(net, Cluster', Cluster_R); if (length(cluster) > ) net = train(net, Cluster', Cluster_R); update_net = ; % Train the neural networks if (min_cluster == ) predicted_output = sim(net, [x;x]); if (min_cluster == ) predicted_output = sim(net, [x;x]); if (min_cluster == ) predicted_output = sim(net, [x;x]); if (min_cluster == ) predicted_output = sim(net, [x;x]);

8 Jones

Metaheuristic Optimization with Evolver, Genocop and OptQuest

Metaheuristic Optimization with Evolver, Genocop and OptQuest Metaheuristic Optimization with Evolver, Genocop and OptQuest MANUEL LAGUNA Graduate School of Business Administration University of Colorado, Boulder, CO 80309-0419 Manuel.Laguna@Colorado.EDU Last revision:

More information

Visual object classification by sparse convolutional neural networks

Visual object classification by sparse convolutional neural networks Visual object classification by sparse convolutional neural networks Alexander Gepperth 1 1- Ruhr-Universität Bochum - Institute for Neural Dynamics Universitätsstraße 150, 44801 Bochum - Germany Abstract.

More information

Opening the Black Box Data Driven Visualizaion of Neural N

Opening the Black Box Data Driven Visualizaion of Neural N Opening the Black Box Data Driven Visualizaion of Neural Networks September 20, 2006 Aritificial Neural Networks Limitations of ANNs Use of Visualization (ANNs) mimic the processes found in biological

More information

Use of Artificial Neural Networks to Investigate the Surface Roughness in CNC Milling Machine

Use of Artificial Neural Networks to Investigate the Surface Roughness in CNC Milling Machine Use of Artificial Neural Networks to Investigate the Surface Roughness in CNC Milling Machine M. Vijay Kumar Reddy 1 1 Department of Mechanical Engineering, Annamacharya Institute of Technology and Sciences,

More information

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting.

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting. An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting. Mohammad Mahmudul Alam Mia, Shovasis Kumar Biswas, Monalisa Chowdhury Urmi, Abubakar

More information

Optimization Methods for Machine Learning (OMML)

Optimization Methods for Machine Learning (OMML) Optimization Methods for Machine Learning (OMML) 2nd lecture Prof. L. Palagi References: 1. Bishop Pattern Recognition and Machine Learning, Springer, 2006 (Chap 1) 2. V. Cherlassky, F. Mulier - Learning

More information

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani Neural Networks CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Biological and artificial neural networks Feed-forward neural networks Single layer

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervised Learning and Clustering Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2009 CS 551, Spring 2009 c 2009, Selim Aksoy (Bilkent University)

More information

Particle Swarm Optimization applied to Pattern Recognition

Particle Swarm Optimization applied to Pattern Recognition Particle Swarm Optimization applied to Pattern Recognition by Abel Mengistu Advisor: Dr. Raheel Ahmad CS Senior Research 2011 Manchester College May, 2011-1 - Table of Contents Introduction... - 3 - Objectives...

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervised Learning and Clustering Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2008 CS 551, Spring 2008 c 2008, Selim Aksoy (Bilkent University)

More information

Application of K-Means Clustering Methodology to Cost Estimation

Application of K-Means Clustering Methodology to Cost Estimation Application of K-Means Clustering Methodology to Cost Estimation Mr. Jacob J. Walzer, Kalman & Co., Inc. 2018 Professional Development & Training Workshop 6/2018 Background There are many ways to analyze

More information

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of

More information

5 Learning hypothesis classes (16 points)

5 Learning hypothesis classes (16 points) 5 Learning hypothesis classes (16 points) Consider a classification problem with two real valued inputs. For each of the following algorithms, specify all of the separators below that it could have generated

More information

HEURISTIC OPTIMIZATION USING COMPUTER SIMULATION: A STUDY OF STAFFING LEVELS IN A PHARMACEUTICAL MANUFACTURING LABORATORY

HEURISTIC OPTIMIZATION USING COMPUTER SIMULATION: A STUDY OF STAFFING LEVELS IN A PHARMACEUTICAL MANUFACTURING LABORATORY Proceedings of the 1998 Winter Simulation Conference D.J. Medeiros, E.F. Watson, J.S. Carson and M.S. Manivannan, eds. HEURISTIC OPTIMIZATION USING COMPUTER SIMULATION: A STUDY OF STAFFING LEVELS IN A

More information

Assignment 4/5 Statistics Due: Nov. 29

Assignment 4/5 Statistics Due: Nov. 29 Assignment 4/5 Statistics 5.301 Due: Nov. 29 1. Two decision rules are given here. Assume they apply to a normally distributed quality characteristic, the control chart has three-sigma control limits,

More information

Clustering Part 4 DBSCAN

Clustering Part 4 DBSCAN Clustering Part 4 Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville DBSCAN DBSCAN is a density based clustering algorithm Density = number of

More information

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 11, November 2014,

More information

Computational Intelligence Meets the NetFlix Prize

Computational Intelligence Meets the NetFlix Prize Computational Intelligence Meets the NetFlix Prize Ryan J. Meuth, Paul Robinette, Donald C. Wunsch II Abstract The NetFlix Prize is a research contest that will award $1 Million to the first group to improve

More information

Analytical model A structure and process for analyzing a dataset. For example, a decision tree is a model for the classification of a dataset.

Analytical model A structure and process for analyzing a dataset. For example, a decision tree is a model for the classification of a dataset. Glossary of data mining terms: Accuracy Accuracy is an important factor in assessing the success of data mining. When applied to data, accuracy refers to the rate of correct values in the data. When applied

More information

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used. 1 4.12 Generalization In back-propagation learning, as many training examples as possible are typically used. It is hoped that the network so designed generalizes well. A network generalizes well when

More information

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X Analysis about Classification Techniques on Categorical Data in Data Mining Assistant Professor P. Meena Department of Computer Science Adhiyaman Arts and Science College for Women Uthangarai, Krishnagiri,

More information

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra Pattern Recall Analysis of the Hopfield Neural Network with a Genetic Algorithm Susmita Mohapatra Department of Computer Science, Utkal University, India Abstract: This paper is focused on the implementation

More information

Optimal Design of a Parallel Beam System with Elastic Supports to Minimize Flexural Response to Harmonic Loading

Optimal Design of a Parallel Beam System with Elastic Supports to Minimize Flexural Response to Harmonic Loading 11 th World Congress on Structural and Multidisciplinary Optimisation 07 th -12 th, June 2015, Sydney Australia Optimal Design of a Parallel Beam System with Elastic Supports to Minimize Flexural Response

More information

Random Search Report An objective look at random search performance for 4 problem sets

Random Search Report An objective look at random search performance for 4 problem sets Random Search Report An objective look at random search performance for 4 problem sets Dudon Wai Georgia Institute of Technology CS 7641: Machine Learning Atlanta, GA dwai3@gatech.edu Abstract: This report

More information

Data Mining with Oracle 10g using Clustering and Classification Algorithms Nhamo Mdzingwa September 25, 2005

Data Mining with Oracle 10g using Clustering and Classification Algorithms Nhamo Mdzingwa September 25, 2005 Data Mining with Oracle 10g using Clustering and Classification Algorithms Nhamo Mdzingwa September 25, 2005 Abstract Deciding on which algorithm to use, in terms of which is the most effective and accurate

More information

University of Florida CISE department Gator Engineering. Clustering Part 4

University of Florida CISE department Gator Engineering. Clustering Part 4 Clustering Part 4 Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville DBSCAN DBSCAN is a density based clustering algorithm Density = number of

More information

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant

More information

Graphical Analysis of Data using Microsoft Excel [2016 Version]

Graphical Analysis of Data using Microsoft Excel [2016 Version] Graphical Analysis of Data using Microsoft Excel [2016 Version] Introduction In several upcoming labs, a primary goal will be to determine the mathematical relationship between two variable physical parameters.

More information

Proceedings of the 2016 International Conference on Industrial Engineering and Operations Management Detroit, Michigan, USA, September 23-25, 2016

Proceedings of the 2016 International Conference on Industrial Engineering and Operations Management Detroit, Michigan, USA, September 23-25, 2016 Neural Network Viscosity Models for Multi-Component Liquid Mixtures Adel Elneihoum, Hesham Alhumade, Ibrahim Alhajri, Walid El Garwi, Ali Elkamel Department of Chemical Engineering, University of Waterloo

More information

Introduction to Pattern Recognition Part II. Selim Aksoy Bilkent University Department of Computer Engineering

Introduction to Pattern Recognition Part II. Selim Aksoy Bilkent University Department of Computer Engineering Introduction to Pattern Recognition Part II Selim Aksoy Bilkent University Department of Computer Engineering saksoy@cs.bilkent.edu.tr RETINA Pattern Recognition Tutorial, Summer 2005 Overview Statistical

More information

Optimal Clustering and Statistical Identification of Defective ICs using I DDQ Testing

Optimal Clustering and Statistical Identification of Defective ICs using I DDQ Testing Optimal Clustering and Statistical Identification of Defective ICs using I DDQ Testing A. Rao +, A.P. Jayasumana * and Y.K. Malaiya* *Colorado State University, Fort Collins, CO 8523 + PalmChip Corporation,

More information

Recent Developments in the Design and Optimization of Constant Force Electrical Contacts

Recent Developments in the Design and Optimization of Constant Force Electrical Contacts Recent Developments in the Design and Optimization of Constant orce Electrical Contacts John C. Meaders * Stephen P. Harston Christopher A. Mattson Brigham Young University Provo, UT, 84602, USA Abstract

More information

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Statistical Analysis of Metabolomics Data Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Outline Introduction Data pre-treatment 1. Normalization 2. Centering,

More information

Using CODEQ to Train Feed-forward Neural Networks

Using CODEQ to Train Feed-forward Neural Networks Using CODEQ to Train Feed-forward Neural Networks Mahamed G. H. Omran 1 and Faisal al-adwani 2 1 Department of Computer Science, Gulf University for Science and Technology, Kuwait, Kuwait omran.m@gust.edu.kw

More information

A HYBRID METHOD FOR SIMULATION FACTOR SCREENING. Hua Shen Hong Wan

A HYBRID METHOD FOR SIMULATION FACTOR SCREENING. Hua Shen Hong Wan Proceedings of the 2006 Winter Simulation Conference L. F. Perrone, F. P. Wieland, J. Liu, B. G. Lawson, D. M. Nicol, and R. M. Fujimoto, eds. A HYBRID METHOD FOR SIMULATION FACTOR SCREENING Hua Shen Hong

More information

Graphical Analysis. Figure 1. Copyright c 1997 by Awi Federgruen. All rights reserved.

Graphical Analysis. Figure 1. Copyright c 1997 by Awi Federgruen. All rights reserved. Graphical Analysis For problems with 2 variables, we can represent each solution as a point in the plane. The Shelby Shelving model (see the readings book or pp.68-69 of the text) is repeated below for

More information

Supporting Information. High-Throughput, Algorithmic Determination of Nanoparticle Structure From Electron Microscopy Images

Supporting Information. High-Throughput, Algorithmic Determination of Nanoparticle Structure From Electron Microscopy Images Supporting Information High-Throughput, Algorithmic Determination of Nanoparticle Structure From Electron Microscopy Images Christine R. Laramy, 1, Keith A. Brown, 2, Matthew N. O Brien, 2 and Chad. A.

More information

CHAPTER 7 MASS LOSS PREDICTION USING ARTIFICIAL NEURAL NETWORK (ANN)

CHAPTER 7 MASS LOSS PREDICTION USING ARTIFICIAL NEURAL NETWORK (ANN) 128 CHAPTER 7 MASS LOSS PREDICTION USING ARTIFICIAL NEURAL NETWORK (ANN) Various mathematical techniques like regression analysis and software tools have helped to develop a model using equation, which

More information

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler BBS654 Data Mining Pinar Duygulu Slides are adapted from Nazli Ikizler 1 Classification Classification systems: Supervised learning Make a rational prediction given evidence There are several methods for

More information

CHAPTER 3 AN OVERVIEW OF DESIGN OF EXPERIMENTS AND RESPONSE SURFACE METHODOLOGY

CHAPTER 3 AN OVERVIEW OF DESIGN OF EXPERIMENTS AND RESPONSE SURFACE METHODOLOGY 23 CHAPTER 3 AN OVERVIEW OF DESIGN OF EXPERIMENTS AND RESPONSE SURFACE METHODOLOGY 3.1 DESIGN OF EXPERIMENTS Design of experiments is a systematic approach for investigation of a system or process. A series

More information

SYDE Winter 2011 Introduction to Pattern Recognition. Clustering

SYDE Winter 2011 Introduction to Pattern Recognition. Clustering SYDE 372 - Winter 2011 Introduction to Pattern Recognition Clustering Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 5 All the approaches we have learned

More information

Numerical Experiments with a Population Shrinking Strategy within a Electromagnetism-like Algorithm

Numerical Experiments with a Population Shrinking Strategy within a Electromagnetism-like Algorithm Numerical Experiments with a Population Shrinking Strategy within a Electromagnetism-like Algorithm Ana Maria A. C. Rocha and Edite M. G. P. Fernandes Abstract This paper extends our previous work done

More information

An Intelligent Technique for Image Compression

An Intelligent Technique for Image Compression An Intelligent Technique for Image Compression Athira Mayadevi Somanathan 1, V. Kalaichelvi 2 1 Dept. Of Electronics and Communications Engineering, BITS Pilani, Dubai, U.A.E. 2 Dept. Of Electronics and

More information

Evaluation of Neural Networks in the Subject of Prognostics As Compared To Linear Regression Model

Evaluation of Neural Networks in the Subject of Prognostics As Compared To Linear Regression Model International Journal of Engineering & Technology IJET-IJENS Vol:10 No:06 50 Evaluation of Neural Networks in the Subject of Prognostics As Compared To Linear Regression Model A. M. Riad, Hamdy K. Elminir,

More information

Lecture 25: Review I

Lecture 25: Review I Lecture 25: Review I Reading: Up to chapter 5 in ISLR. STATS 202: Data mining and analysis Jonathan Taylor 1 / 18 Unsupervised learning In unsupervised learning, all the variables are on equal standing,

More information

Predictive Analytics: Demystifying Current and Emerging Methodologies. Tom Kolde, FCAS, MAAA Linda Brobeck, FCAS, MAAA

Predictive Analytics: Demystifying Current and Emerging Methodologies. Tom Kolde, FCAS, MAAA Linda Brobeck, FCAS, MAAA Predictive Analytics: Demystifying Current and Emerging Methodologies Tom Kolde, FCAS, MAAA Linda Brobeck, FCAS, MAAA May 18, 2017 About the Presenters Tom Kolde, FCAS, MAAA Consulting Actuary Chicago,

More information

APPLICATION OF MULTIPLE RANDOM CENTROID (MRC) BASED K-MEANS CLUSTERING ALGORITHM IN INSURANCE A REVIEW ARTICLE

APPLICATION OF MULTIPLE RANDOM CENTROID (MRC) BASED K-MEANS CLUSTERING ALGORITHM IN INSURANCE A REVIEW ARTICLE APPLICATION OF MULTIPLE RANDOM CENTROID (MRC) BASED K-MEANS CLUSTERING ALGORITHM IN INSURANCE A REVIEW ARTICLE Sundari NallamReddy, Samarandra Behera, Sanjeev Karadagi, Dr. Anantha Desik ABSTRACT: Tata

More information

Enhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques

Enhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques 24 Enhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques Enhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques Ruxandra PETRE

More information

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS Neural Networks Classifier Introduction INPUT: classification data, i.e. it contains an classification (class) attribute. WE also say that the class

More information

Clustering & Classification (chapter 15)

Clustering & Classification (chapter 15) Clustering & Classification (chapter 5) Kai Goebel Bill Cheetham RPI/GE Global Research goebel@cs.rpi.edu cheetham@cs.rpi.edu Outline k-means Fuzzy c-means Mountain Clustering knn Fuzzy knn Hierarchical

More information

Linear Separability. Linear Separability. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons

Linear Separability. Linear Separability. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons Linear Separability Input space in the two-dimensional case (n = ): - - - - - - w =, w =, = - - - - - - w = -, w =, = - - - - - - w = -, w =, = Linear Separability So by varying the weights and the threshold,

More information

Lecture-17: Clustering with K-Means (Contd: DT + Random Forest)

Lecture-17: Clustering with K-Means (Contd: DT + Random Forest) Lecture-17: Clustering with K-Means (Contd: DT + Random Forest) Medha Vidyotma April 24, 2018 1 Contd. Random Forest For Example, if there are 50 scholars who take the measurement of the length of the

More information

Metaheuristic Development Methodology. Fall 2009 Instructor: Dr. Masoud Yaghini

Metaheuristic Development Methodology. Fall 2009 Instructor: Dr. Masoud Yaghini Metaheuristic Development Methodology Fall 2009 Instructor: Dr. Masoud Yaghini Phases and Steps Phases and Steps Phase 1: Understanding Problem Step 1: State the Problem Step 2: Review of Existing Solution

More information

Supervised Learning (contd) Linear Separation. Mausam (based on slides by UW-AI faculty)

Supervised Learning (contd) Linear Separation. Mausam (based on slides by UW-AI faculty) Supervised Learning (contd) Linear Separation Mausam (based on slides by UW-AI faculty) Images as Vectors Binary handwritten characters Treat an image as a highdimensional vector (e.g., by reading pixel

More information

Learning Common Features from fmri Data of Multiple Subjects John Ramish Advised by Prof. Tom Mitchell 8/10/04

Learning Common Features from fmri Data of Multiple Subjects John Ramish Advised by Prof. Tom Mitchell 8/10/04 Learning Common Features from fmri Data of Multiple Subjects John Ramish Advised by Prof. Tom Mitchell 8/10/04 Abstract Functional Magnetic Resonance Imaging (fmri), a brain imaging technique, has allowed

More information

Evolutionary Optimization of Neural Networks for Face Detection

Evolutionary Optimization of Neural Networks for Face Detection Evolutionary Optimization of Neural Networks for Face Detection Stefan Wiegand Christian Igel Uwe Handmann Institut für Neuroinformatik, Ruhr-Universität Bochum, 44780 Bochum, Germany Viisage Technology

More information

Outlier detection using autoencoders

Outlier detection using autoencoders Outlier detection using autoencoders August 19, 2016 Author: Olga Lyudchik Supervisors: Dr. Jean-Roch Vlimant Dr. Maurizio Pierini CERN Non Member State Summer Student Report 2016 Abstract Outlier detection

More information

Unsupervised Learning

Unsupervised Learning Outline Unsupervised Learning Basic concepts K-means algorithm Representation of clusters Hierarchical clustering Distance functions Which clustering algorithm to use? NN Supervised learning vs. unsupervised

More information

A GENETIC ALGORITHM FOR CLUSTERING ON VERY LARGE DATA SETS

A GENETIC ALGORITHM FOR CLUSTERING ON VERY LARGE DATA SETS A GENETIC ALGORITHM FOR CLUSTERING ON VERY LARGE DATA SETS Jim Gasvoda and Qin Ding Department of Computer Science, Pennsylvania State University at Harrisburg, Middletown, PA 17057, USA {jmg289, qding}@psu.edu

More information

DATA MINING AND MACHINE LEARNING. Lecture 6: Data preprocessing and model selection Lecturer: Simone Scardapane

DATA MINING AND MACHINE LEARNING. Lecture 6: Data preprocessing and model selection Lecturer: Simone Scardapane DATA MINING AND MACHINE LEARNING Lecture 6: Data preprocessing and model selection Lecturer: Simone Scardapane Academic Year 2016/2017 Table of contents Data preprocessing Feature normalization Missing

More information

Identification of Multisensor Conversion Characteristic Using Neural Networks

Identification of Multisensor Conversion Characteristic Using Neural Networks Sensors & Transducers 3 by IFSA http://www.sensorsportal.com Identification of Multisensor Conversion Characteristic Using Neural Networks Iryna TURCHENKO and Volodymyr KOCHAN Research Institute of Intelligent

More information

MIS2502: Data Analytics Clustering and Segmentation. Jing Gong

MIS2502: Data Analytics Clustering and Segmentation. Jing Gong MIS2502: Data Analytics Clustering and Segmentation Jing Gong gong@temple.edu http://community.mis.temple.edu/gong What is Cluster Analysis? Grouping data so that elements in a group will be Similar (or

More information

Acoustic to Articulatory Mapping using Memory Based Regression and Trajectory Smoothing

Acoustic to Articulatory Mapping using Memory Based Regression and Trajectory Smoothing Acoustic to Articulatory Mapping using Memory Based Regression and Trajectory Smoothing Samer Al Moubayed Center for Speech Technology, Department of Speech, Music, and Hearing, KTH, Sweden. sameram@kth.se

More information

Unsupervised Learning : Clustering

Unsupervised Learning : Clustering Unsupervised Learning : Clustering Things to be Addressed Traditional Learning Models. Cluster Analysis K-means Clustering Algorithm Drawbacks of traditional clustering algorithms. Clustering as a complex

More information

A Dynamical Systems Analysis of Collaboration Methods in Cooperative Co-evolution.

A Dynamical Systems Analysis of Collaboration Methods in Cooperative Co-evolution. A Dynamical Systems Analysis of Collaboration Methods in Cooperative Co-evolution. Elena Popovici and Kenneth De Jong George Mason University Fairfax, VA epopovic@gmu.edu kdejong@gmu.edu Abstract Cooperative

More information

A Weighted Majority Voting based on Normalized Mutual Information for Cluster Analysis

A Weighted Majority Voting based on Normalized Mutual Information for Cluster Analysis A Weighted Majority Voting based on Normalized Mutual Information for Cluster Analysis Meshal Shutaywi and Nezamoddin N. Kachouie Department of Mathematical Sciences, Florida Institute of Technology Abstract

More information

Design and Performance Analysis of and Gate using Synaptic Inputs for Neural Network Application

Design and Performance Analysis of and Gate using Synaptic Inputs for Neural Network Application IJIRST International Journal for Innovative Research in Science & Technology Volume 1 Issue 12 May 2015 ISSN (online): 2349-6010 Design and Performance Analysis of and Gate using Synaptic Inputs for Neural

More information

INVESTIGATING DATA MINING BY ARTIFICIAL NEURAL NETWORK: A CASE OF REAL ESTATE PROPERTY EVALUATION

INVESTIGATING DATA MINING BY ARTIFICIAL NEURAL NETWORK: A CASE OF REAL ESTATE PROPERTY EVALUATION http:// INVESTIGATING DATA MINING BY ARTIFICIAL NEURAL NETWORK: A CASE OF REAL ESTATE PROPERTY EVALUATION 1 Rajat Pradhan, 2 Satish Kumar 1,2 Dept. of Electronics & Communication Engineering, A.S.E.T.,

More information

Empirical Study on Impact of Developer Collaboration on Source Code

Empirical Study on Impact of Developer Collaboration on Source Code Empirical Study on Impact of Developer Collaboration on Source Code Akshay Chopra University of Waterloo Waterloo, Ontario a22chopr@uwaterloo.ca Parul Verma University of Waterloo Waterloo, Ontario p7verma@uwaterloo.ca

More information

Box-Cox Transformation for Simple Linear Regression

Box-Cox Transformation for Simple Linear Regression Chapter 192 Box-Cox Transformation for Simple Linear Regression Introduction This procedure finds the appropriate Box-Cox power transformation (1964) for a dataset containing a pair of variables that are

More information

1 Case study of SVM (Rob)

1 Case study of SVM (Rob) DRAFT a final version will be posted shortly COS 424: Interacting with Data Lecturer: Rob Schapire and David Blei Lecture # 8 Scribe: Indraneel Mukherjee March 1, 2007 In the previous lecture we saw how

More information

FEATURE EXTRACTION USING FUZZY RULE BASED SYSTEM

FEATURE EXTRACTION USING FUZZY RULE BASED SYSTEM International Journal of Computer Science and Applications, Vol. 5, No. 3, pp 1-8 Technomathematics Research Foundation FEATURE EXTRACTION USING FUZZY RULE BASED SYSTEM NARENDRA S. CHAUDHARI and AVISHEK

More information

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask Machine Learning and Data Mining Clustering (1): Basics Kalev Kask Unsupervised learning Supervised learning Predict target value ( y ) given features ( x ) Unsupervised learning Understand patterns of

More information

Seismic regionalization based on an artificial neural network

Seismic regionalization based on an artificial neural network Seismic regionalization based on an artificial neural network *Jaime García-Pérez 1) and René Riaño 2) 1), 2) Instituto de Ingeniería, UNAM, CU, Coyoacán, México D.F., 014510, Mexico 1) jgap@pumas.ii.unam.mx

More information

Lecture #11: The Perceptron

Lecture #11: The Perceptron Lecture #11: The Perceptron Mat Kallada STAT2450 - Introduction to Data Mining Outline for Today Welcome back! Assignment 3 The Perceptron Learning Method Perceptron Learning Rule Assignment 3 Will be

More information

Supervised Learning in Neural Networks (Part 2)

Supervised Learning in Neural Networks (Part 2) Supervised Learning in Neural Networks (Part 2) Multilayer neural networks (back-propagation training algorithm) The input signals are propagated in a forward direction on a layer-bylayer basis. Learning

More information

Module 7 VIDEO CODING AND MOTION ESTIMATION

Module 7 VIDEO CODING AND MOTION ESTIMATION Module 7 VIDEO CODING AND MOTION ESTIMATION Version ECE IIT, Kharagpur Lesson Block based motion estimation algorithms Version ECE IIT, Kharagpur Lesson Objectives At the end of this less, the students

More information

Box-Cox Transformation

Box-Cox Transformation Chapter 190 Box-Cox Transformation Introduction This procedure finds the appropriate Box-Cox power transformation (1964) for a single batch of data. It is used to modify the distributional shape of a set

More information

Using Excel for Graphical Analysis of Data

Using Excel for Graphical Analysis of Data Using Excel for Graphical Analysis of Data Introduction In several upcoming labs, a primary goal will be to determine the mathematical relationship between two variable physical parameters. Graphs are

More information

Semi-Supervised Clustering with Partial Background Information

Semi-Supervised Clustering with Partial Background Information Semi-Supervised Clustering with Partial Background Information Jing Gao Pang-Ning Tan Haibin Cheng Abstract Incorporating background knowledge into unsupervised clustering algorithms has been the subject

More information

Understanding Competitive Co-evolutionary Dynamics via Fitness Landscapes

Understanding Competitive Co-evolutionary Dynamics via Fitness Landscapes Understanding Competitive Co-evolutionary Dynamics via Fitness Landscapes Elena Popovici and Kenneth De Jong George Mason University, Fairfax, VA 3 epopovic@gmu.edu, kdejong@gmu.edu Abstract. Co-evolutionary

More information

Optimization of Complex Systems with OptQuest

Optimization of Complex Systems with OptQuest Optimization of Complex Systems with OptQuest MANUEL LAGUNA Graduate School of Business Administration University of Colorado, Boulder, CO 80309-0419 Manuel.Laguna@Colorado.EDU Latest revision: April 8,

More information

3 Nonlinear Regression

3 Nonlinear Regression CSC 4 / CSC D / CSC C 3 Sometimes linear models are not sufficient to capture the real-world phenomena, and thus nonlinear models are necessary. In regression, all such models will have the same basic

More information

CS 8520: Artificial Intelligence. Machine Learning 2. Paula Matuszek Fall, CSC 8520 Fall Paula Matuszek

CS 8520: Artificial Intelligence. Machine Learning 2. Paula Matuszek Fall, CSC 8520 Fall Paula Matuszek CS 8520: Artificial Intelligence Machine Learning 2 Paula Matuszek Fall, 2015!1 Regression Classifiers We said earlier that the task of a supervised learning system can be viewed as learning a function

More information

Using neural nets to recognize hand-written digits. Srikumar Ramalingam School of Computing University of Utah

Using neural nets to recognize hand-written digits. Srikumar Ramalingam School of Computing University of Utah Using neural nets to recognize hand-written digits Srikumar Ramalingam School of Computing University of Utah Reference Most of the slides are taken from the first chapter of the online book by Michael

More information

OPTIMIZING A VIDEO PREPROCESSOR FOR OCR. MR IBM Systems Dev Rochester, elopment Division Minnesota

OPTIMIZING A VIDEO PREPROCESSOR FOR OCR. MR IBM Systems Dev Rochester, elopment Division Minnesota OPTIMIZING A VIDEO PREPROCESSOR FOR OCR MR IBM Systems Dev Rochester, elopment Division Minnesota Summary This paper describes how optimal video preprocessor performance can be achieved using a software

More information

Keywords: Extraction, Training, Classification 1. INTRODUCTION 2. EXISTING SYSTEMS

Keywords: Extraction, Training, Classification 1. INTRODUCTION 2. EXISTING SYSTEMS ISSN XXXX XXXX 2017 IJESC Research Article Volume 7 Issue No.5 Forex Detection using Neural Networks in Image Processing Aditya Shettigar 1, Priyank Singal 2 BE Student 1, 2 Department of Computer Engineering

More information

A novel firing rule for training Kohonen selforganising

A novel firing rule for training Kohonen selforganising A novel firing rule for training Kohonen selforganising maps D. T. Pham & A. B. Chan Manufacturing Engineering Centre, School of Engineering, University of Wales Cardiff, P.O. Box 688, Queen's Buildings,

More information

Data Mining. Neural Networks

Data Mining. Neural Networks Data Mining Neural Networks Goals for this Unit Basic understanding of Neural Networks and how they work Ability to use Neural Networks to solve real problems Understand when neural networks may be most

More information

Inital Starting Point Analysis for K-Means Clustering: A Case Study

Inital Starting Point Analysis for K-Means Clustering: A Case Study lemson University TigerPrints Publications School of omputing 3-26 Inital Starting Point Analysis for K-Means lustering: A ase Study Amy Apon lemson University, aapon@clemson.edu Frank Robinson Vanderbilt

More information

Understanding Competitive Co-evolutionary Dynamics via Fitness Landscapes

Understanding Competitive Co-evolutionary Dynamics via Fitness Landscapes Understanding Competitive Co-evolutionary Dynamics via Fitness Landscapes Elena Popovici and Kenneth De Jong George Mason University, Fairfax, VA 3 epopovic@gmu.edu, kdejong@gmu.edu Abstract. Co-evolutionary

More information

Introduction to ANSYS DesignXplorer

Introduction to ANSYS DesignXplorer Lecture 5 Goal Driven Optimization 14. 5 Release Introduction to ANSYS DesignXplorer 1 2013 ANSYS, Inc. September 27, 2013 Goal Driven Optimization (GDO) Goal Driven Optimization (GDO) is a multi objective

More information

SOCIAL MEDIA MINING. Data Mining Essentials

SOCIAL MEDIA MINING. Data Mining Essentials SOCIAL MEDIA MINING Data Mining Essentials Dear instructors/users of these slides: Please feel free to include these slides in your own material, or modify them as you see fit. If you decide to incorporate

More information

COMPUTATIONAL INTELLIGENCE SEW (INTRODUCTION TO MACHINE LEARNING) SS18. Lecture 6: k-nn Cross-validation Regularization

COMPUTATIONAL INTELLIGENCE SEW (INTRODUCTION TO MACHINE LEARNING) SS18. Lecture 6: k-nn Cross-validation Regularization COMPUTATIONAL INTELLIGENCE SEW (INTRODUCTION TO MACHINE LEARNING) SS18 Lecture 6: k-nn Cross-validation Regularization LEARNING METHODS Lazy vs eager learning Eager learning generalizes training data before

More information

EMO A Real-World Application of a Many-Objective Optimisation Complexity Reduction Process

EMO A Real-World Application of a Many-Objective Optimisation Complexity Reduction Process EMO 2013 A Real-World Application of a Many-Objective Optimisation Complexity Reduction Process Robert J. Lygoe, Mark Cary, and Peter J. Fleming 22-March-2013 Contents Introduction Background Process Enhancements

More information

Introduction to Deep Learning

Introduction to Deep Learning ENEE698A : Machine Learning Seminar Introduction to Deep Learning Raviteja Vemulapalli Image credit: [LeCun 1998] Resources Unsupervised feature learning and deep learning (UFLDL) tutorial (http://ufldl.stanford.edu/wiki/index.php/ufldl_tutorial)

More information

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION 6.1 INTRODUCTION Fuzzy logic based computational techniques are becoming increasingly important in the medical image analysis arena. The significant

More information

Tracking Changing Extrema with Particle Swarm Optimizer

Tracking Changing Extrema with Particle Swarm Optimizer Tracking Changing Extrema with Particle Swarm Optimizer Anthony Carlisle Department of Mathematical and Computer Sciences, Huntingdon College antho@huntingdon.edu Abstract The modification of the Particle

More information

A Genetic Algorithm-Based Approach for Building Accurate Decision Trees

A Genetic Algorithm-Based Approach for Building Accurate Decision Trees A Genetic Algorithm-Based Approach for Building Accurate Decision Trees by Z. Fu, Fannie Mae Bruce Golden, University of Maryland S. Lele,, University of Maryland S. Raghavan,, University of Maryland Edward

More information

IMPROVEMENTS TO THE BACKPROPAGATION ALGORITHM

IMPROVEMENTS TO THE BACKPROPAGATION ALGORITHM Annals of the University of Petroşani, Economics, 12(4), 2012, 185-192 185 IMPROVEMENTS TO THE BACKPROPAGATION ALGORITHM MIRCEA PETRINI * ABSTACT: This paper presents some simple techniques to improve

More information