Particle Swarm Optimization Based Learning Method for Process Neural Networks

Similar documents
A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

An Immune Concentration Based Virus Detection Approach Using Particle Swarm Optimization

A Data Classification Algorithm of Internet of Things Based on Neural Network

THREE PHASE FAULT DIAGNOSIS BASED ON RBF NEURAL NETWORK OPTIMIZED BY PSO ALGORITHM

Small World Particle Swarm Optimizer for Global Optimization Problems

Modified Particle Swarm Optimization

An improved PID neural network controller for long time delay systems using particle swarm optimization algorithm

Overlapping Swarm Intelligence for Training Artificial Neural Networks

Improving Tree-Based Classification Rules Using a Particle Swarm Optimization

An Island Based Hybrid Evolutionary Algorithm for Optimization

A Boosting-Based Framework for Self-Similar and Non-linear Internet Traffic Prediction

A Network Intrusion Detection System Architecture Based on Snort and. Computational Intelligence

Using Decision Boundary to Analyze Classifiers

Simultaneous Perturbation Stochastic Approximation Algorithm Combined with Neural Network and Fuzzy Simulation

Learning the Three Factors of a Non-overlapping Multi-camera Network Topology

Experimental Study on Bound Handling Techniques for Multi-Objective Particle Swarm Optimization

Simultaneous Vanishing Point Detection and Camera Calibration from Single Images

Hybrid PSO-SA algorithm for training a Neural Network for Classification

Small World Network Based Dynamic Topology for Particle Swarm Optimization

ANN-Based Modeling for Load and Main Steam Pressure Characteristics of a 600MW Supercritical Power Generating Unit

Constraints in Particle Swarm Optimization of Hidden Markov Models

Feeding the Fish Weight Update Strategies for the Fish School Search Algorithm

Automatic differentiation based for particle swarm optimization steepest descent direction

Performance Assessment of DMOEA-DD with CEC 2009 MOEA Competition Test Instances

Face Alignment Under Various Poses and Expressions

Traffic Signal Control Based On Fuzzy Artificial Neural Networks With Particle Swarm Optimization

Mobile Robot Path Planning in Static Environments using Particle Swarm Optimization

CONTENT ADAPTIVE SCREEN IMAGE SCALING

Image Classification Using Wavelet Coefficients in Low-pass Bands

RoboCup 2014 Rescue Simulation League Team Description. <CSU_Yunlu (China)>

A Linear Approximation Based Method for Noise-Robust and Illumination-Invariant Image Change Detection

A 3D MODEL RETRIEVAL ALGORITHM BASED ON BP- BAGGING

Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India.

Dynamic Analysis of Structures Using Neural Networks

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Cooperative Coevolution using The Brain Storm Optimization Algorithm

OMBP: Optic Modified BackPropagation training algorithm for fast convergence of Feedforward Neural Network

A Robust Image Zero-Watermarking Algorithm Based on DWT and PCA

Research on Evaluation Method of Product Style Semantics Based on Neural Network

Feature Selection for Multi-Class Imbalanced Data Sets Based on Genetic Algorithm

Computer-Aided Diagnosis for Lung Diseases based on Artificial Intelligence: A Review to Comparison of Two- Ways: BP Training and PSO Optimization

An Automatic Timestamp Replanting Algorithm for Panorama Video Surveillance *

Improving Information Retrieval Effectiveness in Peer-to-Peer Networks through Query Piggybacking

Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic Algorithm and Particle Swarm Optimization

Power Load Forecasting Based on ABC-SA Neural Network Model

An Approach to Polygonal Approximation of Digital CurvesBasedonDiscreteParticleSwarmAlgorithm

Effect of the PSO Topologies on the Performance of the PSO-ELM

Time Series Prediction as a Problem of Missing Values: Application to ESTSP2007 and NN3 Competition Benchmarks

Multi-Criterion Optimal Design of Building Simulation Model using Chaos Particle Swarm Optimization

Particle Swarm Optimization

Visual object classification by sparse convolutional neural networks

Improving Generalization of Radial Basis Function Network with Adaptive Multi-Objective Particle Swarm Optimization

Artificial Neural Network and Multi-Response Optimization in Reliability Measurement Approximation and Redundancy Allocation Problem

SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES. Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari

Particle swarm optimization for mobile network design

Flexible-Hybrid Sequential Floating Search in Statistical Feature Selection

Parallel Evaluation of Hopfield Neural Networks

Neural Network Weight Selection Using Genetic Algorithms

Sequences Modeling and Analysis Based on Complex Network

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

Feature weighting using particle swarm optimization for learning vector quantization classifier

Iterative Removing Salt and Pepper Noise based on Neighbourhood Information

BP Neural Network Based On Genetic Algorithm Applied In Text Classification

MATLAB representation of neural network Outline Neural network with single-layer of neurons. Neural network with multiple-layer of neurons.

Radial Basis Function Neural Network Classifier

Application of Or-based Rule Antecedent Fuzzy Neural Networks to Iris Data Classification Problem

Using CODEQ to Train Feed-forward Neural Networks

ARMA MODEL SELECTION USING PARTICLE SWARM OPTIMIZATION AND AIC CRITERIA. Mark S. Voss a b. and Xin Feng.

Parameter Selection of a Support Vector Machine, Based on a Chaotic Particle Swarm Optimization Algorithm

Inertia Weight. v i = ωv i +φ 1 R(0,1)(p i x i )+φ 2 R(0,1)(p g x i ) The new velocity update equation:

Convex combination of adaptive filters for a variable tap-length LMS algorithm

Open Access Research on the Prediction Model of Material Cost Based on Data Mining

Recognition of Human Body Movements Trajectory Based on the Three-dimensional Depth Data

Reversible Image Data Hiding with Local Adaptive Contrast Enhancement

Research Article Path Planning Using a Hybrid Evolutionary Algorithm Based on Tree Structure Encoding

A MULTI-SWARM PARTICLE SWARM OPTIMIZATION WITH LOCAL SEARCH ON MULTI-ROBOT SEARCH SYSTEM

Fault Diagnosis of Power Transformers using Kernel based Extreme Learning Machine with Particle Swarm Optimization

Rainfall forecasting using Nonlinear SVM based on PSO

FAST RANDOMIZED ALGORITHM FOR CIRCLE DETECTION BY EFFICIENT SAMPLING

Structure-adaptive Image Denoising with 3D Collaborative Filtering

Cell-to-switch assignment in. cellular networks. barebones particle swarm optimization

The Application Research of Neural Network in Embedded Intelligent Detection

EFFICIENT ATTRIBUTE REDUCTION ALGORITHM

PARTICLE SWARM OPTIMIZATION FOR MULTIDIMENSIONAL CLUSTERING OF NATURAL LANGUAGE DATA

A Hybrid Fireworks Optimization Method with Differential Evolution Operators

Research on the New Image De-Noising Methodology Based on Neural Network and HMM-Hidden Markov Models

PARTICLE SWARM OPTIMIZATION (PSO)

Artificial Neuron Modelling Based on Wave Shape

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Fingerprint Ridge Distance Estimation: Algorithms and the Performance*

A NEW ROBUST IMAGE WATERMARKING SCHEME BASED ON DWT WITH SVD

A derivative-free trust-region algorithm for reliability-based optimization

Adaptively Choosing Neighbourhood Bests Using Species in a Particle Swarm Optimizer for Multimodal Function Optimization

A Fuzzy C-means Clustering Algorithm Based on Pseudo-nearest-neighbor Intervals for Incomplete Data

An Application of Genetic Algorithm for Auto-body Panel Die-design Case Library Based on Grid

AN NOVEL NEURAL NETWORK TRAINING BASED ON HYBRID DE AND BP

10-701/15-781, Fall 2006, Final

An Optimized Pixel-Wise Weighting Approach For Patch-Based Image Denoising

Binary Differential Evolution Strategies

Optimization of Observation Membership Function By Particle Swarm Method for Enhancing Performances of Speaker Identification

Transcription:

Particle Swarm Optimization Based Learning Method for Process Neural Networks Kun Liu, Ying Tan, and Xingui He Key Laboratory of Machine Perception, Ministry of Education, Peking University, Beijing 100871, China ytan@pku.edu.cn http://www.cil.pku.edu.cn Abstract. This paper proposes a new learning method for process neural networks (PNNs) based on the Gaussian mixture functions and particle swarm optimization (PSO), called PSO-LM. First, the weight functions of the PNNs are specified as the generalized Gaussian mixture functions (GGMFs). Second, a PSO algorithm is used to optimize the parameters, such as the order of GGMFs, the number of hidden neurons, the coefficients, means and variances of Gaussian functions and the thresholds in PNNs. In PSO-LM, the parameter space is transformed from the function space to real number space by using GGMFs. PSO can give a global search in the real number parameter space by avoiding the premature and gradient calculations in back propagation method. According to our analysis and several experiments, PSO-LM can outperform current basis function expansion based learning method (BFE-LM) for PNNs and the classic back propagation neural networks (BPNNs). Keywords: Process Neural Networks, Learning Method, Particle Swarm Optimization, Gaussian Mixture Model. 1 Introduction Recently, He et al. proposed this process neural networks (PNNs) to deal with process inputs, which can be related to time, locations etc. [1]. Inputs such as time series are usually sensitive to the time parameter, so the weights and thresholds should also be related to time in order to accumulate the effects of inputs more precisely. In PNNs, there exist process neurons (PNs). The inputs, connection weights and threshold of a PN can all be functions of time. Theoretically, the approximation capabilities of PNNs are better than classic artificial neural networks (ANNs) when solving for time-varying function problems. ANNs are considered as the special case of PNNs. PNNs have been applied in many actual tasks, such as simulation of oil reservoir exploitation [1] and traffic flow prediction [2]. In this study, we propose this new learning method for PNNs. Instead of transforming the inputs and weight functions, we specify the weight functions connected to the process neurons as the generalized Gaussian mixture functions (GGMFs), which theoretically can approximate any continuous functions. Afterward, instead of using the Corresponding Author. L. Zhang, J. Kwok, and B.-L. Lu (Eds.): ISNN 2010, Part I, LNCS 6063, pp. 280 287, 2010. c Springer-Verlag Berlin Heidelberg 2010

Particle Swarm Optimization Based Learning Method for Process Neural Networks 281 classic back propagation method, we use PSO to optimize the parameters, such as the order and coefficients of GGMFs, the means and variances of Gaussian functions, the number of hidden neurons and their thresholds. This new learning method for PNNs is called PSO-LM for short. The remainder of this paper is organized as follows. Section 2 describes the original PNN model and a standard PSO model. The PSO-LM is proposed in Section 3 with discussions about its qualities. Section 4 elaborates two experiments to compare the PSO-LM with the existing BFE-LM for PNNs and the BPNN. Concluding remarks are drawn in Section 5. 2 Background 2.1 Process Neural Networks PNN is a generalized model of traditional neural networks, the inputs are related to instantaneous conditions [1]. The key of PNNs is the special structure of PNs. Fig. 1(a) shows a simple process neuron model (SPN), in which x i (t)(i =1, 2,...,n) are the input time-varying functions and w i (t)(i =1, 2,...,n) are the corresponding weight functions. K( ) is the aggregate function of the SPN, which can be summation or integration etc.. f( ) is the activation function, with a threshold value θ. The output of the SPN is a constant. A simple feed forward process neural network (PNN) is shown in Fig. 1(b). The current learning methods for PNNs include: basis function expansion method (BFE) [1] and numerical learning method (NL) [3]. The BFE includes two steps: first, transform the input data and the weight functions of the PNN into the space that expanded by some specified orthogonal basis functions. Therefore, there will be no timevarying parameters in the model. Second, adjust the parameters in the networks by back propagation method. However, how to choose appropriate basis functions and how many expansion items should be reserved are awkward to decide. Furthermore, transforming the input data in the first place certainly will lose information contained in the input signals, which can restrain the performance of PNNs. The NL method is only suitable for small-scale discrete input signals. When the length of the input series increases, the number of parameters in NL will increase (a) Simple Process Neuron Model (b) Feed-Forward Process Neural Network Fig. 1. Process Neural Networks

282 K. Liu, Y. Tan, and X. He rapidly, which can lead to huge computational complexity. Furthermore, these two learning methods both rely on the back propagation (BP) method, which will suffer from the premature convergence problem and gradient calculations. 2.2 Particle Swarm Optimization Particle swarm optimization (PSO) [4], as a stochastic global optimization technique based on a social interaction metaphor, can be used to train ANNs effectively [5]. Bratton and Kennedy defined a standard PSO model (SPSO) in [6]. The update rules can be equivalently transformed and expressed as follows: V id (t +1)=ωV id (t)+c 1 ɛ 1 (P ibd (t) X id (t)) +c 2 ɛ 2 (P nbd (t) X id (t)), (1) X id (t +1)=X id (t)+v id (t +1). (2) where i =1, 2,,n, n is the number of particles in the swarm, d =1, 2,,D,and D is the dimensionality of solution space. The learning factors c 1 and c 2 are nonnegative constants, r 1 and r 2 are random numbers uniformly drawn from the interval [0, 1], which are all scalar quantities for each particle in each dimension. P ibd and P nbd are the locations of the best positions found so far by particle i and its neighbors in dimension d. ω is called inertia weight, which determines the effect of the inertia to the movements of particles. 3 The Proposed PSO-LM for PNNs It is impossible to learn the weight functions directly from the function space. In this study, we assume that all weight functions are GGMFs based on the fact that all the continuous functions can be approximated by GGMFs. Afterwards, the constant parameters in the model are optimized by PSO. 3.1 Generalized Gaussian Mixture Functions A standard Gaussian mixture model is defined in (3), and its qualities are expressed in Theorem 1. For the proof of the theorem, one can refer to [7]. f = k a i N(u i,σ i ), (3) i=1 where N(u i,σ i ) is the normal distribution with mean u i and variance σ i, Σ k i=1 a i =1, a i 0,i=1, 2,...,kand k is the order of the model. Theorem 1 (Gaussian Mixture Model Approximation Theorem). Finite Gaussian mixture model can approximate a non-negative Riemann integrable function in real number field by any degree of accuracy. Particularly, it is able to approximate any probability density functions by any degree of accuracy.

Particle Swarm Optimization Based Learning Method for Process Neural Networks 283 The generalized Gaussian mixture function is also defined as in (3), only with the difference that a i R. Therefore, one can derive the following corollary. The proof of this corollary is straight according to the above theorem. Corollary 1. Generalized Gaussian mixture functions can approximate a continuous function in real number field by any degree of accuracy. 3.2 Learning Method for PNNs Based on PSO (PSO-LM) Consider the feed-forward PNN shown in Fig. 1(b). Only the hidden neurons are simple process neurons, and the other parameters are all constants. Theoretically, the approximation capability of this PNN can be retained if the weight functions are specialized as generalized Gaussian mixture functions. There are n input neurons. The following parameters will be optimized by PSO according to specific objectives: the order of GGMFs (k), the number of hidden neurons (m), the coefficients, means and variances of Gaussian functions corresponding to the weight functions that connect from input neurons to hidden neurons (a ijl,u ijl,σ ijl,i= 1,...,n; j = 1,...,m; l = 1,...,k), the thresholds of hidden neurons (θ j,j = 1,...,m) and the output neuron (θ). So, a particle in the swarm of PSO can be expressed as: p =(k, m, a ijl,u ijl,σ ijl,θ j,w j,θ),i=1,...,n; j =1,...,m; l =1,...,k. (4) Each position of the particle in the search space represents one process neural network. The PSO will search for a PNN that has appropriate structure and parameter values to optimize the objective function, which is the performance evaluation of the PNN. When the swarm of PSO converges to one position, which is supposed to be the best parameter setups of the PNN, the learning process terminates. According to the above study, the PSO-LM for PNNs can be summarized as in Algorithm 1: Algorithm 1. PSO-LM for PNNs Step 1: Set up the objective function of the training as the fitness function of PSO and the generalized Gaussian mixture functions as the weight functions of PNs shown in Fig. 1(a). Step 2: Set parameters of PSO (particle number p, w, c, search space borderlines). Step 3: Each particle in PSO is formed according to (4). Step 4: Run PSO. Step 5: Stop criterion (maximum iterations or/and fixed learning precision). Step 6: If the result is not satisfactory, go back to Step 3. 3.3 Discussions The proposed PSO-LM has several advantages comparing with the existing learning methods for PNNs.

284 K. Liu, Y. Tan, and X. He First, PSO-LM does not need any transformation to the input signals, which will not lose any information carried in inputs. BFE-LM extracts the principal components of the input data by using the basis function expansion, which will retain most of the information. However, the detailed information carried in inputs is the most needed information in some task. For example, in pattern recognition, two patterns might have the same first and second principal components, which will be reserved according to BFE-LM. But, it is the detailed differences carried in their third components that can distinguish them, which will be eliminated by BFE-LM. Thereafter, samples from different classes can be misclassified into the same pattern. Furthermore, the transformations also bring more computation. Second, gradient information is no longer needed in PSO-LM. The structure and the parameters of the PNNs can be optimized simultaneously by PSO. The learning capability will be enhanced, which is guaranteed by the strong optimization capability of PSO. Furthermore, PSO-LM is able to solve high dimensional data learning problems by using high dimensional GGMFs as the weight functions. PSO-LM can also solve multiobjectives training problems as PSO is good at doing multi-objective optimization. The computational complexity of PSO-LM for PNNs can be expressed as C = F P I. F is the amount of one time feed-forward calculation of PNN which is also one time fitness calculation for PSO, P is the number of particles in PSO, and I is the maximum number of iterations. The generalization performance of PSO-LM can be verified by the following experiments. 4 Experiments 4.1 Traffic Flow Prediction The data is from PeMS system [8], which is a performance measurement system of freeway traffic from University of California. In this study, we chose traffic volume data recorded every five minutes in the time interval 10:00 to 15:00 during the weekdays from 2009/6/1 to 2009/6/21. The data is taken from five sequential loop detectors, whose serial numbers are 716810, 717907, 717898, 717896 and 717887 in District 7. Experimental Setups. The objective is to predict the traffic flow of loop 717898, which lies in the middle of the five loops, in the next five minutes. We will use the previous five records from all the five loops to predict the next value of the third loop [3]. The data from the first two weeks which contains 120 samples is used to train the three models mentioned above, and the data of the third week which has 55 samples is used as the test data to test their generalization performances. The parameters of the three models are listed in Table 1. The number of db1 wavelet basis functions chosen in BFE-LM is ten, which is decided by trial-and-error and is also the number of neurons in the first hidden layer of BFE-LM. k is the order of the GGMFs and m is the number of neurons in hidden layer of PSO-LM, which will both be decided by PSO to optimize the objective function. Finally, the activation functions of neurons are all chosen to be purelin function according to experiments. Notice that k =2and m =13is decided by PSO.

Particle Swarm Optimization Based Learning Method for Process Neural Networks 285 Table 1. Parameter Setups For Traffic Flow Experiment Structures Activation F unctions P arameters BPNN 5-25-1 purelin, purelin BFE LM 5-10-25-1 purelin, purelin db1 wavelet basis PSO LM 5-m-1 purelin, purelin k=2, m=13 The evaluation of the models depends on the following two targets: MPAE = 1 n y t ŷ t (5) n y t=1 t n t=1 RMSE = (y t ŷ t ) 2 (6) n We will compare the models on each day of the weekdays independently. Each of the models ran 30 times independently, and the averaged results were recorded. Experimental Results. The averaged results are shown in Fig. 2. As can be seen, PSO- LM outperforms BFE-LM and BPNN. According to the variance analysis, the p values of MPAE and RMSE are 0.013 and 0.034, which mens the improvements of the performance by PSO-LM is statistically significant. The PSO-LM can learn the data well and predict the traffic flow more precisely, which indicate that PSO-LM can give PNNs more powerful learning capability and also good generalization capability. PSO-LM spent the most time to achieve the best performance for its global search strategy. BFE-LM consumes the least time, because the basis function expansion for the inputs retains only the principal components of the data, which makes it easier for NNs to learn. Nevertheless, BFE-LM can not guarantee good performances. 4.2 Spam Detection In this experiment, the three models will be compared in identifying spam on a standard Spambase data [9]. There are 4601 samples, each of which has 57 attributes and one MPAE 0.11 0.1 0.09 0.08 0.07 PSO LM BFE LM BPNN RMSE 80 70 60 50 PSO LM BFE LM BPNN 0.06 40 0.05 Mon Tue Wed Thu Fri (a) MPAE 30 Mon Tue Wed Thu Fri (b) RMSE (c) Time Consuming Fig. 2. Experimental Results For Traffic Flow Forecasting

286 K. Liu, Y. Tan, and X. He label. The label denotes whether the e-mail is considered as a spam (1) or not (0). In this study, we will use the first 48 attributes which denote the frequencies of 48 words occurring in the e-mails. Experimental Setups. We use cross-validation in this experiment. First, 601 samples are randomly chosen to join a test data. Then, we use the rest 4000 samples to perform the standard ten times cross-validation to train and validate the models. Finally, the trained models will be tested on the test data to show their generalization capabilities. Each model will run 30 times independently to get statistical results. The parameters in this experiment are shown in Table 2. All the parameters are with the same meanings as in Table 1. Table 2. Parameter Setups For Spambase Experiment Structures Activation F unctions P arameters BPNN 48-15-1 tansig, logsig BFE LM 1-10-5-1 tansig, logsig db1 wavelet basis PSO LM 1-m-1 purelin, logsig k=4, m=16 Experimental Results. The averaged results are shown in Fig. 3. The validation accuracies of the ten times cross-validation are shown in Fig. 3(a). As can be seen, PSO-LM can obviously outperform the other two models, which indicates that the PSO-LM has the best learning capabilities. The test accuracies are shown in Fig. 3(b). Similarly, PSO-LM has the best generalization capability. Time consuming comparison is shown in Fig. 3(c). As the same reason as in the above experiment, the PSO-LM needs more time to gain better performance that the other models can not achieve. Furthermore, the p values of the validation and test accuracy are 0.0486 and 0.0491, which means the performance improvements made by PSO-LM are statistically significant. Validation Accuracy 0.92 0.9 0.88 0.86 0.84 0.82 PSO LM BFE LM BPNN 1 2 3 4 5 6 7 8 9 10 Batch (a) Validation Accuracy Test Accuracy 0.92 0.9 0.88 0.86 0.84 0.82 PSO LM BFE LM BPNN 1 2 3 4 5 6 7 8 9 10 Batch (b) Test Accuracy (c) Time Consuming Fig. 3. Experimental Results For Spam Detection

Particle Swarm Optimization Based Learning Method for Process Neural Networks 287 5 Conclusion This paper proposed a new learning method for process neural networks based on the Gaussian mixture weight functions and particle swarm optimization (PSO-LM). The weight functions were specified as the generalized Gaussian mixture functions. The structureand parametersin themodelwerethenoptimized bypso. According to experiments, PSO-LM had better performance on time-series prediction and pattern recognition than BFE-LM and BPNNs. Although it needed more computations for the global search strategy of PSO, it achieved better performances that other methods can not gain. Acknowledgement This work is supported by National Natural Science Foundation of China (NSFC), under grant number 60875080 and 60673020, and partially financially supported by the Research Fund for the Doctoral Program of Higher Education (RFDP) in China. This work is also in part supported by the National High Technology Research and Development Program of China (863 Program), with grant number 2007AA01Z453. References 1. He, X.G., Xu, S.H.: Process Neural Networks. Science Press, Beijing (2007) 2. He, S., Hu, C., Song, G.J., Xie, K.Q., Sun, Y.Z.: Real-Time Short-Term Traffic Flow Forecasting Based on Process Neural Network. In: Sun, F., Zhang, J., Tan, Y., Cao, J., Yu, W. (eds.) ISNN 2008, Part II. LNCS, vol. 5264, pp. 560 569. Springer, Heidelberg (2008) 3. Wu, T.S., Xie, K.Q., Song, G.J., He, X.G.: Numerical Learning Method for Process Neural Network. In: Yu, W., He, H., Zhang, N. (eds.) ISNN 2009. LNCS, vol. 5551, pp. 670 678. Springer, Heidelberg (2009) 4. Kennedy, J., Eberhart, R.C.: Particle Swarm Optimization. In: Proc. of the IEEE International Conference on Neural Networks 1995, pp. 1942 1948 (1995) 5. Van Den Bergh, F., Engelbrecht, A.P.: Training Product Unit Networks Using Cooperative Particle Swarm Optimization. In: Proc. of the IEEE International Joint Conference on Neural Networks 2001, vol. 1, pp. 126 131 (2001) 6. Bratton, D., Kennedy, J.: Defining a Standard for Particle Swarm Optimization. In: Proc. of the IEEE Swarm Intelligence Symposlum (SIS), pp. 120 127 (2007) 7. Yuan, L.H., Li, Z., Song, J.S.: Probabillty Density Function Approximation Using Gaussian Mixture Model. Radio Communications Technology 33(2), 20 22 (2007) 8. Freeway Performance Measurement System (PeMS), http://pems.eecs.berkeley.edu 9. Asuncion, A., Newman, D.J.: UCI Machine Learning Repository. In: School of Information and Computer Science. University of California, Irvine (2007), http://archive.ics.uci.edu/ml/datasets/spambase