Climate Precipitation Prediction by Neural Network

Similar documents
Efficient Rule Set Generation using K-Map & Rough Set Theory (RST)

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Yuki Osada Andrew Cannon

6. NEURAL NETWORK BASED PATH PLANNING ALGORITHM 6.1 INTRODUCTION

Data Mining. Neural Networks

CS 4510/9010 Applied Machine Learning. Neural Nets. Paula Matuszek Fall copyright Paula Matuszek 2016

Intelligent Methods in Modelling and Simulation of Complex Systems

A Neural Network Model Of Insurance Customer Ratings

CS6220: DATA MINING TECHNIQUES

Character Recognition Using Convolutional Neural Networks

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

ROUGH SETS THEORY AND UNCERTAINTY INTO INFORMATION SYSTEM

Optimizing Number of Hidden Nodes for Artificial Neural Network using Competitive Learning Approach

Multilayer Feed-forward networks

Learning. Learning agents Inductive learning. Neural Networks. Different Learning Scenarios Evaluation

2. Neural network basics

Supervised Learning in Neural Networks (Part 2)

Back propagation Algorithm:

Image Compression: An Artificial Neural Network Approach

Classification with Diffuse or Incomplete Information

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

CHAPTER 7 MASS LOSS PREDICTION USING ARTIFICIAL NEURAL NETWORK (ANN)

Logical Rhythm - Class 3. August 27, 2018

COMPUTATIONAL INTELLIGENCE

Robustness of Selective Desensitization Perceptron Against Irrelevant and Partially Relevant Features in Pattern Classification

11/14/2010 Intelligent Systems and Soft Computing 1

Simulation of Back Propagation Neural Network for Iris Flower Classification

AMOL MUKUND LONDHE, DR.CHELPA LINGAM

Neural Networks In Data Mining

Use of Artificial Neural Networks to Investigate the Surface Roughness in CNC Milling Machine

CS 4510/9010 Applied Machine Learning

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

Data Mining and Analytics

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

Artificial Neural Network based Curve Prediction

Neural Network Classifier for Isolated Character Recognition

Neural Networks CMSC475/675

DEVELOPMENT OF NEURAL NETWORK TRAINING METHODOLOGY FOR MODELING NONLINEAR SYSTEMS WITH APPLICATION TO THE PREDICTION OF THE REFRACTIVE INDEX

An Intelligent Technique for Image Compression

Neural Network and Deep Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina

MODELLING OF ARTIFICIAL NEURAL NETWORK CONTROLLER FOR ELECTRIC DRIVE WITH LINEAR TORQUE LOAD FUNCTION

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting.

Deep Learning. Architecture Design for. Sargur N. Srihari

CSC 578 Neural Networks and Deep Learning

IN recent years, neural networks have attracted considerable attention

For Monday. Read chapter 18, sections Homework:

Neural Network Application Design. Supervised Function Approximation. Supervised Function Approximation. Supervised Function Approximation

Artificial Neural Networks

Neuro-Fuzzy Inverse Forward Models

CHAPTER VI BACK PROPAGATION ALGORITHM

Pattern Classification Algorithms for Face Recognition

Design and Performance Analysis of and Gate using Synaptic Inputs for Neural Network Application

Attribute Reduction using Forward Selection and Relative Reduct Algorithm

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Lecture 17: Neural Networks and Deep Learning. Instructor: Saravanan Thirumuruganathan

II. ARTIFICIAL NEURAL NETWORK

International Journal of Advanced Research in Computer Science and Software Engineering

Classification Lecture Notes cse352. Neural Networks. Professor Anita Wasilewska

Hybrid Method with ANN application for Linear Programming Problem

Artificial neural networks are the paradigm of connectionist systems (connectionism vs. symbolism)

CPSC 340: Machine Learning and Data Mining. Deep Learning Fall 2018

Global Journal of Engineering Science and Research Management

Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks

Research Article Forecasting SPEI and SPI Drought Indices Using the Integrated Artificial Neural Networks

Lecture #17: Autoencoders and Random Forests with R. Mat Kallada Introduction to Data Mining with R

Evaluation of Neural Networks in the Subject of Prognostics As Compared To Linear Regression Model

Rough Sets, Neighborhood Systems, and Granular Computing

Keywords: ANN; network topology; bathymetric model; representability.

EFFICIENT ATTRIBUTE REDUCTION ALGORITHM

EE 589 INTRODUCTION TO ARTIFICIAL NETWORK REPORT OF THE TERM PROJECT REAL TIME ODOR RECOGNATION SYSTEM FATMA ÖZYURT SANCAR

WHAT TYPE OF NEURAL NETWORK IS IDEAL FOR PREDICTIONS OF SOLAR FLARES?

Empirical transfer function determination by. BP 100, Universit de PARIS 6

RIMT IET, Mandi Gobindgarh Abstract - In this paper, analysis the speed of sending message in Healthcare standard 7 with the use of back

Neural Networks. Neural Network. Neural Network. Neural Network 2/21/2008. Andrew Kusiak. Intelligent Systems Laboratory Seamans Center

Cryptography Algorithms using Artificial Neural Network Arijit Ghosh 1 Department of Computer Science St. Xavier s College (Autonomous) Kolkata India

Implementation of FPGA-Based General Purpose Artificial Neural Network

Implementation of a Library for Artificial Neural Networks in C

International Research Journal of Computer Science (IRJCS) ISSN: Issue 09, Volume 4 (September 2017)

Ensembles of Neural Networks for Forecasting of Time Series of Spacecraft Telemetry

Global couplled ocean-atmosphere reanalysis and forecast with pelagic ecosystem

Application of Artificial Neural Network for the Inversion of Electrical Resistivity Data

A Study on the Neural Network Model for Finger Print Recognition

arxiv: v1 [cond-mat.dis-nn] 30 Dec 2018

Transactions on Information and Communications Technologies vol 19, 1997 WIT Press, ISSN

NEURAL NETWORK-BASED SEGMENTATION OF TEXTURES USING GABOR FEATURES

Data mining with Support Vector Machine

International Journal of Advance Engineering and Research Development OPTICAL CHARACTER RECOGNITION USING ARTIFICIAL NEURAL NETWORK

IMPLEMENTATION OF FPGA-BASED ARTIFICIAL NEURAL NETWORK (ANN) FOR FULL ADDER. Research Scholar, IIT Kharagpur.

House Price Prediction Using LSTM

DECISION TREE INDUCTION USING ROUGH SET THEORY COMPARATIVE STUDY

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

Recitation Supplement: Creating a Neural Network for Classification SAS EM December 2, 2002

Performance Analysis of Data Mining Classification Techniques

Software Based Transmission Line Fault Analysis

CMPT 882 Week 3 Summary

Ship Energy Systems Modelling: a Gray-Box approach

Neural Network Approach for Automatic Landuse Classification of Satellite Images: One-Against-Rest and Multi-Class Classifiers

ESTIMATION OF SUBSURFACE QANATS DEPTH BY MULTI LAYER PERCEPTRON NEURAL NETWORK VIA MICROGRAVITY DATA

FUZZY INFERENCE SYSTEMS

Transcription:

Journal of Mathematics and System Science 5 (205) 207-23 doi: 0.7265/259-529/205.05.005 D DAVID PUBLISHING Juliana Aparecida Anochi, Haroldo Fraga de Campos Velho 2. Applied Computing Graduate Program, National Institute for Space Research (INPE), Brazil. 2. Associated Laboratory for Computing and Applied Mathematics, National Institute for Space Research (INPE). Received: December 08, 204 / Accepted: January 0, 205 / Published: May 25, 205. Abstract: In this work a neural network model for climate forecasting is presented. The model is built by training a neural network with available reanalysis data. In order to assess the model, the development methodology considers the use of data reduction strategies that eliminate data redundancy thus reducing the complexity of the models. The results presented in this paper considered the use of Rough Sets Theory principles in extracting relevant information from the available data to achieve the reduction of redundancy among the variables used for forecasting purposes. The paper presents results of climate prediction made with the use of the neural network based model. The results obtained in the conducted experiments show the effectiveness of the methodology, presenting estimates similar to observations. Key words: Climate Prediction, Neural Networks, Rough Sets Theory. Introduction The development of regional climate models from data considers the hypothesis that it is possible to extract information from historical data on the behavior of climatic conditions. Accordingly, development methodology needs to have a large number of data to ensure that the model considers a wide range of situations. However, despite able to ensure, in principle, greater robustness to the derived models, the handling of large volumes of data may require much computational. This problem is even more complicated when dealing with weather forecasting tasks that require higher accuracy once it deals directly with every-day life of the population. In order to cope with the demand for computing performance, it is necessary to assemble supercomputers or clusters of powerful computers that may work together to produce numerical results that are inputs for forecasting tasks. Nevertheless, it is not only the computing power that counts in a climate forecasting task. In these regards, it would be very Corresponding author: Juliana Aparecida Anochi, Applied Computing Graduate Program, National Institute for Space Research (INPE), Brazil. E-mail: juliana.anochi@gmail.com. profitable if computational models were constructed based on the existing knowledge. This would permit forecasters to run the climate model for a several different initial conditions thus leading to a variety of different scenarios to be analyzed to support their decision on the prediction of the state of the atmosphere. In order to achieve a computational model that may satisfy the stated conditions, we proposed the use of artificial neural networks to be trained with existing climate data. In this paper considered the use of Rough Sets Theory principles in extracting relevant information from the available data to achieve the reduction among the variables used for forecasting purposes. The data reduction approach allows the derivation of smaller data sets from the resulting reducts for the training phase of the neural network without losing data expressiveness for forecasting purposes The advantage of the methodology presented in this paper is the possibility to derive climate forecasting models from simplified data sets gather from observations and to have a means to make prediction using simplified non-complex computational models based on neural networks. An additional advantage of

208 the approach is the possibility to derive a model without a need for deep understanding of the physics related to atmosphere processes. Artificial Neural Networks (ANN) have emerged as excellent tools for deriving data oriented models because of their inherent characteristic of plasticity that permits the adaptation of the learning task when data is provided. In this regards, the use of an ANN is adequate to derive the forecasting model proposed in this paper. For the problem of building prediction models, as proposed in this paper, are used Multilayer Perceptron and models of recurrent Elman and Jordan [2]. 2. Rough Sets Theory Rough Sets Theory (RST) was proposed in 982 by Zdzislaw Pawlak as a mathematical theory to treat uncertain and imprecise information, by deriving approximations of a data set [7]. The RST is based on the similarities among objects measured by an indiscernibility relation, which establish that a set of objects are similar (indiscernible) if they hold the same values for all of their attributes. Rough sets have been proposed for a very wide variety of applications. In particular, the rough set approach seems to be important for Artificial Intelligence, especially in machine learning, knowledge discovery, data mining, expert systems, approximate reasoning and pattern recognition. In this paper, the RST was used to extract relevant information from meteorological data to perform climate prediction, from a reduced data set. 2. Information Systems Rough Sets Theory uses the concept of Information Systems (IS) in which the available data are represented in a table in which the objects are displayed in the rows and the attributes in the columns [6]. More formally, it is a pair SI = ( U, A), where U is non-empty finite set of objects called the universe and A is a non-empty finite set of attributes, such that a : U Va, for every a A. The set V a is called the value set of a. A decision system (DS) is any information system of the form DS = ( U, A { d}), where d A is the decision attribute. 2.2 Indiscernibility The indiscernibility relation is used as a measure of similarity among objects. Thus, a set of objects with the same attributes are indiscernible if only if their attributes hold the same values from their corresponding domains [5]. This is equivalence relation that may be used to treat problems as redundancy of attributes or the existence of irrelevant attributes in the data assigned to only one representative of a class. For given information system SI = ( U, A) for any set B A there exists an equivalence relation IND A ( B), as defined in Equation, that establish two objects x and x' in U are indiscernible from each other by attributes from B. 2 IND A ( B) = {( x, x ) U a B, a( x) = a( x )} () 2.3 Attribute Reduction The attribute reduction procedure is performed by the discernibility function f A = (B) derived from the discernibility matrix which is a symmetric matrix constructed by comparing the attribute values that discern the objects. The attribute representing discernible values are inserted into the matrix. Each entry in the matrix consists of a set of attributes that distinguish a pair of objects x i and x j as it is stated in M = { a B a( x ) a( x )} (2) i, j i j where i, j n and n = U / IND ( B) A. The discernibility function f A = (B) for a subset B is constructed by concatenating the subsets M i,j = {a * a M j,j } of attributes in each position (cell) of the discernibility matrix M, through a Boolean function of m variables a * *,, a m (corresponding to the attributes) defined as below. The function determines the

209 minimum set of attributes that distinguish any class among the existing classes and it is defined as [7]: * { C j i n C = φ} * f ( a, *, am ) = ij, (3) ij 3. Artificial Neural Network Artificial Neural Networks are computational techniques that present a mathematical model inspired on the neural structure of biological organisms, acquiring knowledge through experience. The network has parallel and distributed processing units or neurons, which calculates certain mathematical functions, typically nonlinear. These neurons can be divided into one or more hidden layers interconnected by synaptic weights (connections), which store the knowledge represented in the model, and serve to balance the input received by each network neuron [2]. The artificial neuron model proposed by McCullock and Pitts in 943, interprets the functioning of the neuron as a simple binary circuit that combines multiple inputs and one output signal. The mathematical description resulted in a model with n input terminals representing the dendrites, and only one output simulating the axon. In the mathematical terms, a neuron k could be described considering the following pair of equations [2]: k n v = w j= kj x i (4) network is the ability to learn from the environment, improving its performance through the learning. A neural network learns from its environment through an interactive process of adjustments applied to its synaptic weights and bias levels [2]. In the context of NN, the learning process can be defined as a set of well defined rules for solving a specific problem. 3.. Multilayer Perceptron The architectures of multilayer perceptron neural models are the most used and known. Typically, this architecture consists of a set of units forming an input layer, one or more intermediate (or hidden) layer(s) of computational units, and an output layer. It is a supervised neural network. A very popular algorithm for training MLP-NN is called backpropagation of error. The backpropagation learning error consists of two steps through the different layers of the network: forward and backward steps. For the forward step, an activity pattern (input vector) is applied thought the nodes of the network, and its effect propagates on the entire the network, layer by layer. Finally, a set of outputs is produced as the actual response of the network. During the backward step, the synaptic weights are all adjusted in accordance with an error correction rule. The response of the network is subtracted from a desired output to produce an error signal. Figure shows the architecture of a multilayer perceptron. y = ϕ ( v + bk ) (5) k k where x, x 2,...,x n are the input signals; w kj are the synaptic weights of neuron k computed during the learning phase; v k is the linear combination among the input signals; b k is the bias; ϕ is the activation function; and y k is the output signal of the neuron. The use of bias b k has the effect of applying an affine transformation to the output v k of the linear combiner in the model. The property of primary significance for a neural Fig. Network multilayer perceptron with an input layer, a hidden layer, and an output layer.

20 3.2. Recurrent Neural Network The recurrent networks are neural networks with one or more feedback loops. Given a multilayer perceptron as the basic building block, the application of global feedback can take a variety of forms. We may have feedback from the output neurons to the input layer. Another possibility is the link among the hidden neurons with the input layer. When the MLP has two or more hidden layers, the possible forms of global feedback expand even further [3]. The recurrent networks Elman and Jordan consists of four layers: an input layer, composed of storage neurons that receive an external signal to propagate it, without modification; a hidden layer where activation functions may be linear or non-linear; a context layer is used to store the activation of neurons; the intermediate layer (it can be used with time delay), and an output layer, neurons whose outputs are linear sums of their respective input signals). 4. Results and Discussion In the dimensionality reduction process the relevant attributes are those that mostly occur in the data, in terms of the indiscernibility relation. The data consists of monthly means from January 2000 to December 2009. The analysis region comprises a selected subregion with coordinates (Lat 22.5S, 35S) to (Lon 60W, 45W). The data was downloaded from the reanalysis data repository from National Center for Environmental Prediction & National Center for Atmospheric Research (NCEP/NCAR). The world is divided into parallels (latitudes) and meridians (longitude). The global data reanalysis grid uses a set with horizontal resolution of 2.5 x 2.5 (latitude x longitude) [4]. In order to evaluate the performance of the models, the mean square error is used: Error = N N ( y ) k d k k= 2 (6) where N is the number of patterns in the data set, y k is the true observational value, and d k is the estimation computed by neural model. The available variables are: air temperature (airt), zonal wind components at vertical levels: 300hPa (v300), 500hPa (v500) and 850hPa (v850), meridional wind components at vertical levels: 300hPa (u300), 500hPa (u500) and 850hPa (u850), surface pressure (spres), specific humidity (shum) and precipitation (prec). The variables that were reduced by the use of RST are: airt (77%), u300 (79%), u850 (80%), v300 (77%), v500 (79%), v850 (70%) and spres (78%). It is to be noticed that the variables that form the reducts have a presence greater than 70% in the dicernibility function. It is observed that of 7 variables out of 0, were considered relevant for South. Figure 2 shows the results for season winter of 2007 in the South region of Brazil. Figure 2-a is the observed precipitation in winter 2007; Figure 2-b corresponds to network Jordan trained with all the available variables in the database; Figure 2-c is the estimation produced by the Elman trained with all the variables; and Figure 2-d corresponds to the estimation produced by the MLP network trained with all the available variables. From the results, the Elman network trained with all the available variables in database presented the best performance than the others networks. Figure 3-a is the observed precipitation in winter 2007; Figure 3-b corresponds to the result produced by the recurrent network Jordan trained with the reduced set of variables processed by rough set theory; Figure 3-c shows the result obtained by the network recurrent Elman that was trained with the reduced set of variables processed by RST; and Figure 3-d corresponds to the estimation produced by the MLP network trained with the reduced set by RST. From the results, the Elman network trained with the set reduced by RST present the better performance than the models trained with all available variables. Indeed, both models ANN-Elman and ANN-MLP computed by RST show a high fidelity with real dynamics.

2 Fig. 2 (a) observed precipitation; (b) ANN-Jordan; (c) ANN-Elman; and (d) MLP network.

22 Fig. 3 (a) Observed precipitation; (b) NN-Jordan; (c) NN-Elman; and (d) NN- MLP.

23 In addition, a significant performance was obtained when the training data was processed to eliminate the existing redundancy in the data. The results show that it is worth to search for hidden information in the data in a way that less (computational) efforts are necessary to construct efficient forecasting models. 5. Conclusions In this paper a methodology was used for data reduction, applied in the study of seasonal weather patterns, using the theory of rough sets (data reduction) and neural networks (forecast) in the derivation of predictive models in order to estimate the state future precipitation variable, in south Brazil. The proposed methodology works well with Elman-ANN using RST. The data reduction approach allows the derivation of smaller data sets derived from the resulting reducts for the training phase of the neural network without losing data expressiveness for forecasting purposes. Thus, the RST technique used in the data reduction process allows the identification of relevant information of the data for climate prediction. In addition, it is a technique that approaches the problem of dealing with huge amounts of data which is a characteristic of the processes in meteorology. Acknowledgment The authors thank CNPq and CAPES for the physical and financial support provided to this research. References [] L. Elman, Finding Structure in Time, Cognitive Science. 4 (990) 79-2. [2] S. Haykin, 998, Neural networks principles and practices, Prentice Hall. [3] L. Fausett, 993, Fundamentals of Neural Networks: Architectures, Algorithms And Applications, Prentice-Hall. [4] E. Kalnay, 2003, Atmospheric modeling, data assimilation, and predictability. Cambridge university press. [5] J. Komorowski, Z. Pawlak, L. Polkowski, A. Skowron, Rough sets: A tutorial. Rough fuzzy hybridization: A new trend in decision-making. 999, pp. 3-98. [6] J. Komorowski. A. Ohrn, Modelling prognostic power of cardiac tests using rough sets, Artificial Intelligence in Medicine, 5 (999) 67-9. [7] Z. Pawlak, Rough sets, International Journal of Computer and Information Sciences, (982) 34-356.