ANN Based Short Term Load Forecasting Paradigms for WAPDA Pakistan

Similar documents
Planar Robot Arm Performance: Analysis with Feedforward Neural Networks

CHAPTER VI BACK PROPAGATION ALGORITHM

Gdansk University of Technology Faculty of Electrical and Control Engineering Department of Control Systems Engineering

A neural network that classifies glass either as window or non-window depending on the glass chemistry.

Comparison and Evaluation of Artificial Neural Network (ANN) Training Algorithms in Predicting Soil Type Classification

A Study of Various Training Algorithms on Neural Network for Angle based Triangular Problem

Image Compression: An Artificial Neural Network Approach

The Pennsylvania State University. The Graduate School DEVELOPMENT OF AN ARTIFICIAL NEURAL NETWORK FOR DUAL LATERAL HORIZONTAL WELLS IN GAS RESERVOIRS

Artificial Neural Network (ANN) Approach for Predicting Friction Coefficient of Roller Burnishing AL6061

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Assignment # 5. Farrukh Jabeen Due Date: November 2, Neural Networks: Backpropation

Research on Evaluation Method of Product Style Semantics Based on Neural Network

Early tube leak detection system for steam boiler at KEV power plant

PERFORMANCE COMPARISON OF BACK PROPAGATION AND RADIAL BASIS FUNCTION WITH MOVING AVERAGE FILTERING AND WAVELET DENOISING ON FETAL ECG EXTRACTION

Liquefaction Analysis in 3D based on Neural Network Algorithm

Neuro-Fuzzy Computing

A NEURAL NETWORK BASED IRIS RECOGNITION SYSTEM FOR PERSONAL IDENTIFICATION

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS

Supervised Learning in Neural Networks (Part 2)

The Pennsylvania State University. The Graduate School. John and Willie Leone Family Department of Energy and Mineral Engineering

Dynamic Analysis of Structures Using Neural Networks

Multi Layer Perceptron trained by Quasi Newton learning rule

Neural Network model for a biped robot

An Evaluation of Statistical Models for Programmatic TV Bid Clearance Predictions

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting.

Load forecasting of G.B. Pant University of Agriculture & Technology, Pantnagar using Artificial Neural Network

A Framework of Hyperspectral Image Compression using Neural Networks

MLPQNA-LEMON Multi Layer Perceptron neural network trained by Quasi Newton or Levenberg-Marquardt optimization algorithms

INVESTIGATING DATA MINING BY ARTIFICIAL NEURAL NETWORK: A CASE OF REAL ESTATE PROPERTY EVALUATION

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Proceedings of the 2016 International Conference on Industrial Engineering and Operations Management Detroit, Michigan, USA, September 23-25, 2016

NN-GVEIN: Neural Network-Based Modeling of Velocity Profile inside Graft-To-Vein Connection

CHAPTER 7 MASS LOSS PREDICTION USING ARTIFICIAL NEURAL NETWORK (ANN)

Opening the Black Box Data Driven Visualizaion of Neural N

Using Artificial Neural Networks for Prediction Of Dynamic Human Motion

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

International Journal of Electrical and Computer Engineering 4: Application of Neural Network in User Authentication for Smart Home System

1 The Options and Structures in the Neural Net

Learning. Learning agents Inductive learning. Neural Networks. Different Learning Scenarios Evaluation

Dr. Qadri Hamarsheh Supervised Learning in Neural Networks (Part 1) learning algorithm Δwkj wkj Theoretically practically

International Research Journal of Computer Science (IRJCS) ISSN: Issue 09, Volume 4 (September 2017)

Artificial Neural Network Methodology for Modelling and Forecasting Maize Crop Yield

Solar Radiation Data Modeling with a Novel Surface Fitting Approach

11/14/2010 Intelligent Systems and Soft Computing 1

What is machine learning?

CHAPTER 8 COMPOUND CHARACTER RECOGNITION USING VARIOUS MODELS

Data Mining. Neural Networks

Perceptrons and Backpropagation. Fabio Zachert Cognitive Modelling WiSe 2014/15

Lecture 20: Neural Networks for NLP. Zubin Pahuja

Feed Forward Neural Network for Solid Waste Image Classification

Artificial Neural Network-Based Prediction of Human Posture

Multilayer Feed-forward networks

THE NEURAL NETWORKS: APPLICATION AND OPTIMIZATION APPLICATION OF LEVENBERG-MARQUARDT ALGORITHM FOR TIFINAGH CHARACTER RECOGNITION

MATLAB representation of neural network Outline Neural network with single-layer of neurons. Neural network with multiple-layer of neurons.

Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India.

CHAPTER 5 NEURAL NETWORK BASED CLASSIFICATION OF ELECTROGASTROGRAM SIGNALS

Application of Artificial Neural Network for the Inversion of Electrical Resistivity Data

6. NEURAL NETWORK BASED PATH PLANNING ALGORITHM 6.1 INTRODUCTION

Classification Lecture Notes cse352. Neural Networks. Professor Anita Wasilewska

Neural Networks Laboratory EE 329 A

Accelerating the convergence speed of neural networks learning methods using least squares

PERFORMANCE ANALYSIS AND VALIDATION OF CLUSTERING ALGORITHMS USING SOFT COMPUTING TECHNIQUES

Optimizing Number of Hidden Nodes for Artificial Neural Network using Competitive Learning Approach

Visual object classification by sparse convolutional neural networks

Spatial Variation of Sea-Level Sea level reconstruction

1. Approximation and Prediction Problems

Neural network based Numerical digits Recognization using NNT in Matlab

Artificial neural networks are the paradigm of connectionist systems (connectionism vs. symbolism)

Role of Hidden Neurons in an Elman Recurrent Neural Network in Classification of Cavitation Signals

Prediction of False Twist Textured Yarn Properties by Artificial Neural Network Methodology

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions

Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks

Character Recognition Using Convolutional Neural Networks

Computational Intelligence Meets the NetFlix Prize

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers

MULTILAYER PERCEPTRON WITH ADAPTIVE ACTIVATION FUNCTIONS CHINMAY RANE. Presented to the Faculty of Graduate School of

Neural Network Neurons

Week 3: Perceptron and Multi-layer Perceptron

Performance Evaluation of Artificial Neural Networks for Spatial Data Analysis

AN NOVEL NEURAL NETWORK TRAINING BASED ON HYBRID DE AND BP

Neural Networks. Lab 3: Multi layer perceptrons. Nonlinear regression and prediction.

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra

ALGORITHMS FOR INITIALIZATION OF NEURAL NETWORK WEIGHTS

Automatic Adaptation of Learning Rate for Backpropagation Neural Networks

Fuzzy Mod. Department of Electrical Engineering and Computer Science University of California, Berkeley, CA Generalized Neural Networks

An Improved Backpropagation Method with Adaptive Learning Rate

International Journal of Advanced Research in Computer Science and Software Engineering

Volume 1, Issue 3 (2013) ISSN International Journal of Advance Research and Innovation

The Use of Artificial Neural Networks in Predicting Vertical Displacements of Structures

J. Electrical Systems 13-2 (2017): Regular paper

Evaluation of Neural Networks in the Subject of Prognostics As Compared To Linear Regression Model

The Mathematics Behind Neural Networks

A Compensatory Wavelet Neuron Model

CMPT 882 Week 3 Summary

ISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013

CS6220: DATA MINING TECHNIQUES

Adaptive Regularization. in Neural Network Filters

Hybrid Method with ANN application for Linear Programming Problem

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION

A Data Classification Algorithm of Internet of Things Based on Neural Network

Transcription:

Australian Journal of Basic and Applied Sciences, 4(5): 932-947, 2010 ISSN 1991-8178 ANN Based Short Term Load Forecasting Paradigms for WAPDA Pakistan Laiq Khan, Kamran Javed, Sidra Mumtaz Department of Electrical Engineering, COMSATS Institute of Information Technology, Abbottabad, Pakistan. Abstract: This paper discusses a comprehensive approach for Short Term Load Forecasting (STLF) for Water and Power Development Authority (WAPDA), Pakistan. Keeping in view, non-linear power systems load characteristics; two different Artificial Neural Network (ANN) based architectures have been devised for STLF. Proposed architectures were trained and tested using previous five years actual load data obtained from WAPDA i.e., year 1991-95. Several back propagation based training algorithms were applied to both architectures. Data engineering approach was applied for thorough comparison and identification of the most appropriate approach that yields accurate forecast. Key words: Short term load forecasting, artificial neural network, back propagation, data engineering. INTRODUCTION One of the major research areas in the field of electrical engineering which has been of great importance for many years is load forecasting. It performs as a core process for effective economic planning as well as operational decisions of electric utility companies. Accurate forecast provides basis for decisions in unit commitment, hydrothermal coordination, hydro scheduling, fuel allocation, economic load dispatch and optimal power flow calculations, which leads to savings for electric utilities. Load forecasting facilitates to maintain a balance between electricity supply and demand. Besides this, system operator utilizes the forecasted results for off-line network analysis (Gross and Galiana, 1987). In terms of planning short term load forecasting (STLF) having prediction period of one day or a week, provides basis for economic purposes (Yalcinoz, T., 2005). STLF facilitates with necessary information on the basis of power system characteristics to perform daily operational tasks economically. It plays an important role for secure operational strategies and economic optimization. The aim of STLF is to provide basic generation scheduling functions, corrective actions based on future conditions for off-line security and timely dispatcher information. Forecasting errors result in the form of increased operational costs. Under prediction of STLF leads to a failure for providing the required reserves, thus resulting in increased operational costs by using costly peaking units. On the other hand large reserve capacity and high operational costs are caused due to over prediction of STLF. Thus, lower operational costs due to accurate forecasts yield benefits in the form of savings passed on to the customers (Al-Shareef, A.J., 2008; Sanjib Mishra, 2008). Various approaches have been applied to the problem of STLF. They may be classified into two types, such as statistical approaches and Artificial Intelligence (AI) based techniques. Statistical methods have limited ability and lack of flexibility to model complexity and nonlinearity. On the other hand AI methods are more flexible as far as complexity and nonlinearity are concerned (Temraz, H.K., 1996; Mcmenamin, J.S. and F.A. Monforte, 1998). Statistical methods include time series (Ibrahim Moghram, 1989; Lru, K., 1996), similar day approach (Eugene A. Feinberg,) and regression methods (Alex, D., C. Timothy, 1990), whereas ANN modeling approach is related to AI (Wasserman, P.D., 1989). This paper proposes a novel approach for STLF based on different ANN modeling architectures. As ANN has capability to approximate nonlinear functions, different training sets of historical load can be utilized as input variables. When trained ANN model is presented with unseen data, due to good approximation capability it can generalize among different training sets as well as produce corresponding output. Fig. 1.1 shows patterns of hourly mean load from actual load data obtained from WAPDA for year 1991-95 and 2000-04. Corresponding Author: Laiq Khan, Department of Electrical Engineering, COMSATS Institute of Information Technology, Abbottabad, Pakistan. E-mail: laiq@ciit.net.pk, 932

Fig. 1: Hourly mean load (1991-95, 2000-04). This paper is organized as follows. Section 2 presents an overview of Artificial Neural Network and different types of back propagation algorithms. Section 3 highlights importance of Data Engineering and proposed ANN based architectures. Section 4 deals with simulation results discussion and interactive Graphical User Interface (GUI). Finally Section 5 concludes this research work. 2. Artificial Neural Network: 2.1 ANN Architecture: The concept of Artificial Neural Network was introduced by McCulloch and Pitts in year 1943. ANN techniques having strong learning ability and good performance on unseen data i.e., generalization can achieve function approximation, classification and pattern recognition etc. ANN is inspired from biological neurons, so neurons provide the information processing ability. ANN provides the layered approach consisting of three layers namely input layer, hidden layer and output layer. The input layer collects the input, computations are performed by the hidden layer and then the output of the network is produced by the output layer. Each hidden node as well as output node applies an activation function to its net input value. An activation function can be linear or non-linear, as shown in Fig. 2. Neurons are information processing elements arranged in parallel for performing computational tasks. Connected paths among various neurons provide means for information which also determines the functionality of the network. Each connection provides the weighted strength, in addition a bias is also a weight and its input is usually 1. For a given data set, adaptive learning of ANN is achieved by adjustment of these weights among different elements of network, comparison between network output and desired output (Lee, K.Y., J.H. Park, 1992). When there exists an intermediate layer of neurons between input and output layer, with no feedback links, then the structure is said to be multilayer feed forward network (MFNN) as shown in Fig. 3. MFFN can have more than one hidden layers for processing. 2.2 Back Propagation Algorithm: Back propagation algorithm is a supervised learning method and is of great importance for neural network. It is generalization of Widrow-Hoff rule (1986). It updates weights of ANN model and minimizes mean square error between actual and desired output. Back propagation algorithm is given as follows: While stopping condition is false For each input vector r,1r R 933

Fig. 2: (a) and (b) show structure plots of linear and nonlinear activation function. Fig. 3: Multi layer feed forward network. Calculate hidden neuron inputs 1 1,0 Net r, j wt ji, xri, Calculate hidden neuron outputs 1 1,0 j i x s wt x r, j, r, i Calculate inputs to the output neurons 2 2,1 1 Net wt r,k k, j xr, j Calculate the network outputs 2,1 1 or,k s wtk, j xr, j Calculate the error between o r,k and desired output d rk, Update the weights between hidden an output neurons 2,1 2 1 wt d o s ' Net x k, j rk, rk, rk, r, j (1) 934

Update the weights between hidden an output neurons ' 1,0 2,1 ' wt d o s Net wt s Net x End For End While 2 1 1 j, i r, k r, k r, k k, j r, j r, j (2) 2.3 Different Back Propagation Algorithms: There are a variety of back propagation algorithms based on two things, weight up-gradation for minimizing the value of error and variations in learning rate to achieve convergence (Lahiri, S.K., K.C. Ghanta, 2008). Fig. 4 shows structure of different back propagation algorithms. Fig. 4: Structure of different back propagation algorithms. 2.3.1 Gradient Decent with Momentum: This algorithm functions like a low pass filter as it produces smoothing effect which facilitates the network with the ability to neglect small features that exist in error surface. It has the ability to overcome the problem of local minima. Its functionality is dependant upon the values of learning rate and the momentum constant. 2.3.2 Batch Gradient Decent with Variable Learning Rate: The enhancement in performance of back propagation algorithm can be achieved by changing learning rate during training session, hence increasing the level of complexity of error surface. Higher values of learning rate causes instability and very small values of learning rate causes increased time factor for convergence. Changes in the values of weights and biases are proportional to learning rate multiplied by the value of negative gradient. For achieving adaptive learning rate there must be some replacements in training methodology used by gradient decent algorithm. Initially, calculation of network output and error values, after that the value of current learning rate is utilized for calculating new values of weights and biases for every epoch. Then the values of output and error are calculated and if the resulting value of error exceeds previously acquired value of error the learning rate is decreased and vice versa. Higher values of learning rate can lead to instability. 935

2.3.3 Resilient Back Propagation: The value of slope in sigmoid functions that are used for multilayer neural networks must minimize to zero value for increased values of input. So, in case of steep decent algorithm this factor causes a problem, as the values of gradient can attain small magnitude and have a very minimum effect on the values of weights and biases. The purpose of this algorithm is to overcome such phenomena. Therefore, as the direction of the weight update is determined by its sign, so there is no effect produced by the magnitude of the derivative. The change in values of weights and biases is dependant upon three conditions i.e., if for two successive iterations the derivative of the performance function has same sign with respect to changes in weight, then the up-gradation values of biases and weights are increased. Otherwise the values the up- gradation of biases and weights are decreased for opposite scenario, and for zero value of derivative the up gradation value remains unchanged. 2.3.4 Conjugate Gradient Algorithms: The values of weights in back propagation algorithm are adjusted by negative of gradient, thus showing the direction of performance function decrease. So this approach does not acquire the fast convergence. Therefore, faster convergence is acquired by conjugate gradient algorithms when search is based on the conjugate directions. Fletcher Reeves update, Polka Ribiere update and Powell Beale restarts are the examples of Conjugate gradient algorithm. 2.3.5 Quasi newton Algorithm: It is an alternative approach for fast convergence and optimized outcome rather than conjugate gradient algorithm and do not need calculation of second derivatives. For each iteration, up-gradation of an approximate Hessian matrix is achieved. This update is acquired as a function of gradient. BFGS algorithm and Levenberg Marquardt algorithm are examples of Quasi-Newton algorithm. 2.3.6 One Step Secant Algorithm: This algorithm facilitates in reduction of gap as well as creating a bridge between quasi Newton algorithms and conjugate gradient algorithm. This algorithm has two main advantages i.e., it does not keep Hessian matrix obtained at each iteration and also considers previously obtained Hessian matrix equivalent to identity matrix. Its edge over other algorithms is that it can calculate new search directions without calculating a matrix inverse. 3. Proposed ANN Architectures: 3.1 Data Engineering and Modeling Strategy: Data engineering is basically an approach for changing data into important information and matching it with the models (Olaf Wolkenhauer, 2001). It is based on the blend of Pattern Recognition approaches with System Theory. Fig. 5 shows the involvement of different levels towards data engineering approach that can lead to prediction. Fig. 5: Data engineering methodology. Many methodologies have been devised for the purpose of load forecasting using ANN approach [17-18]. In this work, ANN based model development regarding STLF was met by using neural network toolbox. For ANN based STLF architecture modeling, sequence of steps was involved, that are mentioned below: 1. Preprocessing hourly load dataset. 2. Neural Network (NN) architecture modeling. 3. Training/ Testing of NN model. 4. Post-training analysis of forecast model. 3.1.2 Pre-processing Load Dataset: Pre-processing of five years dataset was performed prior to training and testing. Input/output dataset was normalized to lie between 0 and 1. Normalized dataset was divided into training and testing datasets. 936

3.1.3 ANN Architecture Modeling: Two different ANN architectures were proposed for a day ahead forecast, namely a Multilayer Feed Forward Network (MFFN) and an ANN engine consisting of twenty four such MFFN architectures. Network topology of MFFN consisted of a three layered structure namely input layer, hidden layer and output layer. The specific load at a given hour does not simply depend upon its previous hour load, but also the actual load at the same hour of previous day, week as well as month. So, keeping in view such technicalities, architectures modeled for a day ahead forecast are shown in Fig. 6 (a) and (b). (a) MFFN for STLF Fig. 6: Different architectures for day ahead forecast. (b) ANN engine for STLF 937

3.1.4 Training and Testing of ANN Model: Each forecasting model was evaluated after training. Training dataset consisted of four years of hourly load dataset from year 1991-94. Different types of back propagation based methodologies shown in Fig. 4 were utilized for training as well as evaluating performance of forecasting model. Log-sigmoid activation function was applied to hidden layer neurons, where as linear activation function was utilized for output layer neurons. As far as testing scenario is concerned, both architectures were evaluated upon hourly load dataset of year 1995. Parameters relating forecasting models are summarized in Table. 1. Table 1: ANN parameters ANN-Parameters Settings Number of input layer neurons 3 Number of hidden layer neurons 5 Number of output layer neurons 1 Activation function for hidden layer Logsig Activation function for output layer Pureline Epochs 100 Performance function MSE Training data 4 years Test data 1 year 3.5 Post-training Analysis: For measuring the accuracy of forecast, different performance measures were applied to the output of the model for measuring the forecast accuracy which are mentioned below: Over fitting condition: Used to check whether data is being memorized or not. Condition for overfitting: if v> v + v else Nooverfitting i.e., (validation error > average validation error + standard deviation of validation error). Curve fitting toolbox was used to check the network s efficiency and goodness of fit based on following criterions: Sum of squares due to error (SSE). n SSE= w (y -y ) i=1 R-square. n i=1 i i i SSR= w i( yi-y i) i=1 n SST= w ( y -y ) ^ 2 i i i Root mean squared error (RMSE). MSE y 2 2 RMSE = Where as, ^ i represents target, yi represents model output and y i represents average target value. The quality of forecasting was checked by different error criterions. Mean square error. 938

MSE= yt 2 pred -yt act N Mean absolute percentage error. MAPE y y 100 tpred tact y N y tact y Where as, tact represents actual output and tpred represents predicted output of the model. RESULTS AND DISCUSSION 4.1 Comparative Analysis: To evaluate the performance of ANN based STLF models, different criterions have been devised to identify the specific technique for optimal solution to STLF problem. Therefore, to analyze results in a better way, comparison is done on the basis of applied data engineering approaches and different performance measures. The comparative results obtained from various training methodologies are shown in Table 2. Table 2: Outcomes of STLF models (A) MFFN forecast model. # ANN Algo N/W topology H-Act Ftn O/p Act Ftn Epochs Time (s) MSE MAPE Goodness of fit ------------------- ---------------------------------------------------------------------------------------------- I L HL OL Correlation Over-fitting SSE RSQR RMSE Coefficient 1 trainb 3 5 1 logsig Purelin 100 28.9 0.0015 5.265 0.9342 NO 0.0356 0.0031 0.0385 2 trainbfg 3 5 1 logsig Purelin 100 67.56 0.0002 2.6482 0.9575 NO 0.0057 0.0113 0.0154 3 trainbr 3 5 1 logsig Purelin 100 92.1 0.0002 2.4289 0.9594 NO 0.0049 0.01 0.0144 4 traincgb 3 5 1 logsig Purelin 100 92.1 0.0003 2.7623 0.9538 NO 0.0062 0.0122 0.0161 5 traincgf 3 5 1 logsig Purelin 100 50.1 0.0002 2.3869 0.9639 NO 0.0054 0.0068 0.015 6 traincgp 3 5 1 logsig Purelin 100 41.04 0.0002 2.7264 0.9463 NO 0.0091 0.0091 0.0156 7 traingd 3 5 1 logsig Purelin 100 29.755 0.0012 5.7993 0.7743 NO 0.029 0.0015 0.0347 8 traingda 3 5 1 logsig Purelin 100 32.86 0.0005 3.4104 0.9097 NO 0.012 0.0137 0.0223 9 traingdm 3 5 1 logsig Purelin 100 29.39 0.0009 4.9159 0.9161 NO 0.0208 0.0024 0.0294 10 traingdx 3 5 1 logsig Purelin 100 34.47 0.0008 4.4216 0.9056 NO 0.0193 0.0021 0.0284 11 trainlm 3 5 1 logsig Purelin 100 92.21 0.0002 2.4659 0.9593 NO 0.005 0.0099 0.0144 12 trainoss 3 5 1 logsig Purelin 100 102.3 0.0003 2.887 0.9442 NO 0.0063 0.0107 0.0163 13 trainrp 3 5 1 logsig Purelin 100 42.79 0.0003 2.9257 0.9392 NO 0.0081 0.0132 0.0183 14 trainscg 3 5 1 logsig Purelin 100 73.4 0.0003 2.9698 0.9551 NO 0.0068 0.0118 0.0168 (B) ANN engine with 24 MFFN # ANN Algo N/W topology H-Act Ftn O/p Act Ftn Epochs Time (s) MSE MAPE Goodness of fit ------------------- ---------------------------------------------------------------------------------------------- I L HL OL Correlation Over-fitting SSE RSQR RMSE Coefficient 1 trainbr 3 5 1 logsig Purelin 100 104.8 0.0002 2.5152 0.9621 NO 0.0049 0.0105 0.0142 2 traincgf 3 5 1 logsig Purelin 100 101.89 0.0004 3.5818 0.9371 NO 0.01 0.012 0.0204 3 trainlm 3 5 1 logsig Purelin 100 166.02 0.0002 2.5086 0.9659 NO 0.0048 0.0105 0.0141 4.2 Speed and Memory Comparison: This comparison is done to ensure that, which algorithm produces fastest output response, therefore, minimum utilization of memory. Because, these factors are dependant upon nature of the problem, network topology, training set etc. From Table 2 (a), training algorithms that require minimum training time are traingd and traingdm with training time 28.9 sec and 29.3sec. 4.3 Network Accuracy: As network accuracy is an important factor to be considered, that is related to different performance measures. Therefore, to measure the performance of network response with certain type of network topology as well as training period, forecasting errors are calculated using MSE and MAPE. From the results shown in Table 2 (a) and (b), training algorithms with more accurate response are traincgf, trainbr and trainlm, with MAPE values 2.3%, 2.4%, 2.4% and MSE values 0.0002 each i.e., for MFFN model. Comparative plots for MAPE and MSE relating both architectures are shown below in Fig. 7. 939

(A) MSE relating MFFN (B) MSE relating ANN engine 940

(C) MAPE relating MFFN Fig. 7: Performance measures for STLF models. (D) MAPE relating ANN engine 4.4 Goodness of Fit (GOF): Different algorithms were utilized to train ANN; none of these resulted over fitting. From the results shown in Table 2 (a), the training algorithms with highest correlation are trainbr, traincgf and trainlm with correlation coefficient values (R-value) 0.959, 0.963, 0.95 and smaller values for other parameters that relate to goodness of fit. 941

R-value is computed by performing linear regression between network target and forecasted output. Fig.8 (a-f) shows comparison between actual and forecasted values of electric load using training algorithms with highest correlation values for both architectures. Comparative plots to analyze goodness of fit for both architectures are shown below in Fig. 9 (a) and (b). 4.5 Observations: Keeping in view different performance criterians, best results were acquired by three training algorithms, which are trainbr, traincgf and trainlm i.e., for MFFN model. These training algorithms were further utilized to train ANN engine (with 24 MFFN). Even with more complex network structure as well as training methodology, ANN engine produced inferior outcomes as compared to MFFN model, as shown in Table 2 (b). Another aspect related to ANN engine was that it took more time for training. Fig. 10 (a) and (b) shows comparative plots relating proposed STLF models considering training algorithms that produced optimum outcomes. (a) trainbr (b) traincgf (c) trainlm (d) trainbr 942

(e) traincgf (f) trainlm Fig. 8: Comparison between target Vs forecasted electric load using MFFN i.e., (a-c) and ANN engine (d-f). (a) GOF for MFFN 943

Fig. 9: Comparative plots relating Goodness of fit. (b) GOF for ANN engine (a) 24 hr forecast for MFFN 944

Fig. 10: Comparative plots for Day ahead forecast (b) 24 hr forecast for ANN engine Further, its clear from the observations that best outcomes were acquired using MFFN model trained with Levenberg Marquart algorithm. 4.6 Graphical User Interface and Program Structure: The front-end Graphical User Interface (GUI) for STLF models provides an interactive and user friendly environment. It facilitates the user with different pop-up menus to modify neural network topology, activation function, training algorithm and period of forecast. After execution of the program via load button, resultant plots are displayed in an arranged manner. This provides the user with good understanding of the forecasting model. Layout of GUI is shown in Fig. 11. This software also permits the user to analyze the predicted output upon different criterions for measuring forecast accuracy. Fig. 12 shows interface to analyze the neural network response through different performance indexes as described in section 3. 5. Conclusion This problem study incorporates the Short Term Load Forecasting (STLF) using two different strategies for ANN modeling. Out of various back propagation training methodologies some are used for the first time to resolve the problem of STLF. Out of the used techniques faster training algorithms for feed forward training furnished more promising results with superior accuracy rather than traditional methodologies devised for STLF problem on account of different data engineering methods and performance indexes. One of the important aspects of this research work is that the performance of different STLF models is evaluated for improved forecast quality using actual load data of five years and secondly, these STLF models do not involve much complexity i.e., network topology as well as limited training period to avoid over fitting and acquiring optimum results. Lastly, an enhanced and interactive Graphical User Interface provides user friendly environment as well as better understanding and analysis ability of different STLF models via front-end interface. 945

Fig. 11: Interface for artificial neural network based load forecasting model. Fig. 12: Interface for performance issues Notations: wt wt x r 1,0 2,1 Hidden layer input weights Hidden layer output weights Input vector 946

1 x d rk, 1 Net 2 Net o rk, s s ' Hidden neuron output Desired output Hidden neuron inputs Inputs to output neurons Network outputs Sigmoid function Sigmoid function (differential) Learning rate REFERENCES Alex, D., C. Timothy, 1990. "A regression based approach to short term load forecasting", IEEE transaction on power systems, 5(4): 1535-1550. Al-Shareef, A.J., E.A. Muhammad and E. Al-Judaibi, 2008. "One Hour Ahead Load Forecasting Using Artificial Neural Network for the Western Area of Saudi Arabia", International Journal of Electrical Systems Science and Engineering, 37: 219-224. Eugene A. Feinberg, Dora Genethliou, "Applied mathematics for power systems Load Forecasting". Gross and Galiana, 1987. "Short-Term Load Forecasting", proceedings of the IEEE, 75(12): 1558-1570. Ibrahim Moghram, Saifure Rahman, 1989. "Analysis Evaluation of five short term load forecasting techniques", IEEE transaction on power systems, 4(4): 1484-1491. Lahiri, S.K., K.C. Ghanta, 2008. "Development of an artificial neural network correlation for prediction of hold-up of slurry transport in pipelines", chemical engineering science, 63: 1497-1509. Lru, K., S., Subbarayan, R.R. Shoults, M.T. Manry, C. Kwan, F.L. Lewis, J. Naccarino, 1996. "Comparison of very short term load forecasting techniques", IEEE transactions on power systems, 11(2): 1-8. Wasserman, P.D., 1989. "Neural computing; Theory and Practice", Van Nostrand Reinhold. Lee, K.Y., J.H. Park, 1992. "Short term load forecasting using artificial neural network", IEEE transactions on power systems, 7(1): 124-131. Mcmenamin, J.S. and F.A. Monforte, 1998. "Short term energy forecasting with neural networks", Energy J., 19(4): 43-61. Olaf Wolkenhauer, 2001. "Data Engineering: Fuzzy Mathematics in Systems Theory and Data Analysis, year. Sanjib Mishra, 2008. "Short Term Load Forecasting Using Computational Intelligence methods", Masters thesis, National Institute of Technology Rourkela. Temraz, H.K., M.M.A. Salama and Quintana, 1996. "Application of the decomposition technique for forecasting the load of a large electric power network", IEEE Proc. Gener. Transm. Distrib., 143(1): 13-18. Widrow, B. and M.E. Hoff, 1986. "Adaptive switching networks", Parallel distributed processing, Cambridge, MA: MIT Press, 1: 123-134. Yalcinoz, T., 2005. "An educational software package for power systems analysis and operation", International Journal of Electrical Engineering Education, 42(4): 369-382. 947