×
files/journal/2022-09-01_23-34-07-000000_997.jpg

Journal of Modern Mathematics and Statistics

ISSN: Online
ISSN: Print 1994-5388
160
Views
0
Downloads

Comparative Analysis of Rainfall Prediction Using Statistical Neural Network and Classical Linear Regression Model

G.N. Amahia and C.G. Udomboso
Page: 66-70 | Received 21 Sep 2022, Published online: 21 Sep 2022

Full Text Reference XML File PDF File

Abstract

Different types of models have been used in modeling rainfall. Since 1990s however, interest has shifted from traditional models to ANN in rainfall modeling. Many researchers found out that the ANN performed better than such traditional models. In this study, we compared a traditional linear model and ANN in the modeling of rainfall in Ibadan, Nigeria. Ibadan is a city in West Africa, located in the tropical rainforest zone, using the data obtained from the Nigeria Meteorological (NIMET) station. Three variables were considered in this study rainfall, temperature and humidity. In selecting between the two models, we concentrated on the choice of adjusted , Akaike Information Criterion (AIC) and Schwarz Information Criterion (SIC). Though, the MSE and R2 were also used, it was concluded from results that MSE is not a good choice for model selection. This is due to the nature of the rainfall data (which has wide variations). It was found that the Statistical Neural Network (SNN), generally performed better than the traditional (OLS).


INTRODUCTION

Rainfall models play an important role in water resource management planning and therefore, different types of models with various degrees of complexity have been developed for this purpose. Researchers have classified these models into three broad categories, regardless of their structural diversity. They are black box or system theoretical models, conceptual models and physically-based models. Black box models normally contain no physically-based input and output transfer functions and therefore are considered to be purely empirical models. Conceptual rainfall-runoff models usually incorporate interconnected physical elements with simplified forms and each element is used to represent a significant or dominant constituent hydrologic process of the rainfall-runoff transformation. Conceptual rainfall models have been widely employed in hydrological modeling.

There has been a tremendous growth in the interest of application of the Artificial Neural Networks (ANNs) in rainfall modeling sinc, the 1990s. ANNs are usually assumed to be powerful tools for functional relationship establishment or nonlinear mapping in various applications. Cannon and Whitfield (2002) found ANNs to be superior to linear regression procedures. Shamseldin (1997) examined the effectiveness of rainfall-runoff modeling with ANNs by comparing their results with the Simple Linear Model (SLM), the seasonally based Linear Perturbation Model (LPM) and the Nearest Neighbor Linear Perturbation Model (NNLPM) and concluded that ANNs could provide more accurate discharge forecasts than some of the traditional models. Presently, more and more researchers are utilizing artificial neural networks because these models possess desirable attributes of universal approximation and the ability to learn from examples. Artificial neural networks constitute a useful tool to predict and forecast various hydrological variables and are used extensively in water resources research (Tayfur, 2002; Cigizoglu, 2003a, b, 2004; Sudheer, 2005; Cigizoglu and Kisi, 2006; Toprak and Cigizoglu, 2008).

The artificial neural network models are frequently employed for rainfall forecasting (Hsu et al., 1997; Kuligowski and Barros, 1998; Hall et al., 1999; Silverman and Dracup, 2000; Applequist et al., 2002; Ramirez et al., 2005; Freiwan and Cigizoglu, 2005). French et al. (1992) used a neural network to forecast rainfall intensity fields in space and whilst Raman and Sunilkumar (1995) used the artificial neural network to synthesize reservoir inflow series for two sites in the Bharathapuzha basin, South India. According to Minns and Hall (1996), majority of the early study in this area have been mainly theoretical, concentrating on neural network performance with artificially generated rainfall-runoff data. ANN concept is a mapping technique. The network maps values from the input to the output by means of a function, generally known as the transfer function. This study seeks to compare traditional linear model and ANN in the modeling of rainfall in Ibadan, Nigeria. Ibadan is a city in West Africa, located in the tropical rainforest zone. It has two distinct climatic seasons; the wet season from May to November and the dry season from December to April. The peak of the rainfall season is usually August/September. The data used in this study are daily observations obtained from the Nigeria Meteorological (NIMET) station in Ibadan for 33 years (1971-2003).

MATERIALS AND METHODS

Prediction using regression model: The regression model is basically partitioned into two parts, namely the predicted portion having the characteristic that can be ascribed to all the observations considered as a group in a parametric framework. The remaining portion called the residual is the difference between the observed and the predicted values which is ascribed to unknown sources. The model is given as:

(1)

where, n is the number of observations, yi is the ith observation, xj = (x1i, x2i, ..., xki) is the predictor variable vector related to yi, β = (β0, β1, ..., βp) is the parameter vector and ei is the error associated with ith observation. The matrix form of Eq. 1 is given as:

(2)

Using least squares, the estimate of the parameter β is derived as:

(3)

The predicted model becomes:

(4)

so that the residual is given as:

(5)

Prediction using SNN model: The Statistical Neural Network (SNN), like the regression model is composed of two parts; the predictive and the residual. It is given as:

(6)

Where:

Thus, Eq. 6 can be written as:

(7)

where, X = (x0, x1, ..., xI) is the vector of the input variable, g(.) is the transfer (or activation) function and w = (α, β, γ) are the weights (or parameters) associated with the input vector, hidden neuron and the transfer function, respectively while ei is the error associated with the network. We note that when there is no hidden neuron, the SNN reduces to the ordinary regression model. The weights are estimated using Taylor’s 1st order approximation:

(8)

Where:

if θ = w - w0 and;

then, we can write Eq. 6 as:

(9)

Where:

Using the Least Squares Method (LCM), the estimate of the parameter θ becomes:

(10)

The estimated model is;

while the network error is given as:

The transfer function used in this study is the symmetric saturated linear transfer function:

The model formulation is 2-2-1, 2-5-1, 2-10-1, 2-50-1, 2-100-1.

Standardization: We standardized the input variables (converting them to the range (0, 1)) before feeding them into the network. Without this standardization, large values input into an ANN would require extremely small weighting factors to be applied. Otherwise:

Due to inaccuracies introduced by floating point calculations on microcomputers, one should avoid using the very small weighting values that would be required
Without using extremely small initial weights, changes made by the backpropagation would be insignificantly small and training would be very sluggish as the gradient of the transfer function at extreme values would be approximately zero. It is this gradient that is used in the adjustment of weights and biases in an ANN during training

All values leaving the network are automatically output in a standardized format. However, these output values must be destandardized to provide meaningful results. This is done by simply reversing the standardization algorithm used on the input nodes. While E-views 4 and SPSS was employed for the OLS part of the analysis, a neural code was written for the analysis of the SNN using Matlab R2009a and interesting results were obtained.

Model selection criteria: In this study, we discuss several criteria that have been used to choose between the two models. Several criteria are used for this purpose. In particular, we discuss these criteria: R2, adjusted , Akaike Information Criterion (AIC) and Schwarz Information Criterion (SIC). All these criteria aim at minimizing the residual sum of squares (SSE). However, except for the 1st criterion, criteria R2, and SIC impose a penalty for including an increasingly large number of predictors. Thus, there is a tradeoff between goodness of fit of the model and its complexity (as judged by the number of predictors). The variables included in the study include rainfall, temperature and humidity. A close observation of the data shows that there is a wide range of variation, especially in the rainfall data. This is expected as rain does not fall every day. The data were further aggregated as monthly data. Rainfall is the predicted variable while temperature and humidity are the predictor variables. The entire data were split into three, n1-n3 where, n2 = 2 n1 and n3 = 3n1 = n1+n2.

RESULTS AND DISCUSSION

Figure 1 compares the time plot for the variables under study. Except for temperature that seems to be stable, rainfall and humidity shows a lot of irregularity. It can be seen from the irregular pattern that there is high humidity when there is low rainfall. The pattern exhibited by the three variables is cyclical.

It is however, noticed that while the MSE for the OLS reduces as the sample increases, it is on the contrary for that of SNN (Table 1). However, generally as the hidden neuron increases, the MSE for the SNN becomes reduced. This is explained by the sensitivity of the neural network to data. Discrepancies not captured in the traditional method affects the network at very low hidden neurons. Increasing the number of hidden neuron reduces the biases in the weights. This explains the reason for the low values of the MSE in higher neurons.

Table 1: Model selection based on OLS and SNN

 

Fig. 1: Time plot for rainfall, temperature and humidity

 

 

Table 2: Model selection based on MSE and R2

 

 

Table 3: Model selection based on AIC and SIC

 

It is important to note here that the MSE is not a good criteria for model selection (Table 2). The model performances of the two models (OLS and SNN) is compared using the adjusted , Akaike Information Criterion (AIC) and Schwarz Information Criterion (SIC) (Table 3). The R2 shows the performances of the individual models. We therefore, notice that at higher neurons, the better the SNN model. The AIC and SIC results show that the SNN a better model to the traditional OLS. However, SNN models for n2 and n3 at 100 hidden neuron was not good enough.

CONCLUSION

We have compared the ordinary least squares regression and the statistical neural network to estimate rainfall events in Ibadan, Nigeria from 1971-2003. Both methods attempt to minimize the error sum of squares between observations and predicted values. Regression requires an explicit function to be defined before the least squares parameter estimates could be computed while a neural network depends more on training data and the learning algorithm. We have restricted the variables for the models to rainfall, temperature and humidity as measured by the Nigeria Meteorological (NIMET) station in Ibadan. Comparing model prediction in both cases show that the statistical neural network performs better than the regression model. Researchers note here that if we consider other variables like sunshine hour, wind speed and solar radiation, the neural network is likely to have an added advantage over the traditional OLC.

ACKNOWLEDGEMENT

Researchers are deeply indebted to Professor Isaac Kwame Dontwi of the Department of Mathematical Sciences of the Kwame Nkrumah University of Science and Technology, Kumasi, Ghana for the scholarly review of this research.

How to cite this article:

G.N. Amahia and C.G. Udomboso. Comparative Analysis of Rainfall Prediction Using Statistical Neural Network and Classical Linear Regression Model.
DOI: https://doi.org/10.36478/jmmstat.2011.66.70
URL: https://www.makhillpublications.co/view-article/1994-5388/jmmstat.2011.66.70