Groundwater-Level Prediction Using Multiple Linear Regression and Artificial Neural Network Techniques: A Comparative Assessment
Groundwater-Level Prediction Using Multiple Linear Regression and Artificial Neural Network Techniques: A Comparative Assessment
net/publication/260724302
CITATIONS READS
170 2,286
2 authors, including:
SEE PROFILE
All content following this page was uploaded by Sasmita Sahoo on 28 April 2015.
Kochi Prefecture
Source: http://wikitravel.org/en/File:Shikoku.png
Study Area
Fig. 1 Location of the study area and observed sites in the Konan groundwater basin
observations of each of the independent variables X1, Hence, the pragmatic form of the statistical regression
X2,…, Xk respectively; β0, β1, β2........, βk are fixed but model obtained after applying the least-square technique
unknown parameters; and εi is a random variable that is is given as (Makridakis et al. 2008):
normally distributed.
The task of MLR modeling is to estimate the unknown
parameters (β0, β1, β2........, βk) of an MLR model (Eq. 1). Y i ¼ b0 þ b1 X l;i þ b2 X 2;i þ ⋯⋯⋯⋯ þ bk X k;i þ ei ð2Þ
bj
t¼ ð6Þ
SE b jÞ
.
SSR p MSR
F¼ . ¼ ð7Þ
Fig. 2 Correlograms for water levels at site A-2 for various time lags, SSE ðn – p – 1Þ MSE
showing: a autocorrelation function; b partial-autocorrelation function
X 2
Yb i −Y SSR Design of ANN models
R2 ¼ X 2 ¼ ð8Þ
SST A neural network is made up of a number of nodes, which
Y i −Y are the processing elements of the network and are usually
called ‘neurons’. Each neuron is connected to other neurons,
receives an input signal, processes it and transforms it into an
where Yi =observed value of a dependent variable, Yb i = output signal (Haykin 1994). In order to execute the function
estimated or predicted value of Y, and Y = mean value of a of biological neurons artificially, a proper design of artificial
dependent variable. neural network (ANN) is necessary. Hence, designing ANN
(or ANN models) follows a systematic procedure involving
n–1 five basics steps: (1) selection of influential inputs; (2)
Adjusted R ¼ 1−
2
1−R2
n–p–1 selection of suitable ANN architecture; (3) building the
0 . 1 neural network; (4) training and testing of the developed
SSE ðn – p – 1Þ ANN models; and (5) performance evaluation of the ANN
¼ 1−@ . A ð9Þ models. The first step is described in section “Study sites and
SST ðn – 1Þ data sources” and a brief description of the remaining steps is
given in the following.
x0 (Bias)
x1 (Rainfall)
x2 (Temperature)
Hidden Layer
x3 (River stage)
x4 (Dummy 1) z0
Output Layer
x5 (Dummy 2) .
.
.
.
. .
Output ( )
x14 (Dummy 11)
. Groundwater-level
.
x15 (Rainfall t-1) .
Output Node (m) = 1
x16
. Connection
(Rainfall t-2) zl
Weight
layer. Thus, the prediction of the ANN model (yk) can be simulated values are as close as possible to the known output
expressed as follows (Haykin 1994): (observed) values. The most common method used for
training feed-forward neural networks is back-propagation
X
l
yk ¼ wjk z j þ b j ð11Þ training (Hagan and Menhaj 1994; Hagan et al. 1996), which
j¼1
is a two-step approach. In the first step, the input signal is
propagated forward to estimate the outputs. In the second step,
where wjk is the connection weight between jth node of
a backward step is performed to adjust the weight vectors
hidden layer and output node k; l is the number of neurons
between the layers to minimize the network error. Although
in the hidden layer; zj is the output of the jth hidden
various weight-optimization techniques are available for the
neuron resulting from all the input data and bj is the
training of ANNs, the Levenberg-Marquardt (LM) algorithm
connection weight for bias term.
was used in this study, owing to its ability to achieve
convergence quickly (Moré 1978; Press et al. 1992). The
Training algorithm LM algorithm trains neural networks at a rate 10 to 100 times
After selecting the ANN architecture, the next step is the faster than the usual gradient-descent back-propagation
training of the network. The purpose of training is to ensure algorithm, and is more effective at finding optimal results
that the network extracts the fundamental characteristics or than standard back-propagation (Hagan and Menhaj 1994).
pattern from the datasets used in ANN modeling. The training The LM algorithm combines the Gauss-Newton algorithm
process consists of determining the ANN weights and biases. and the method of gradient descent, and it uses an
The ANN is trained with a set of input and known output data approximation of the Hessian matrix to update weights as
to determine the connection weights, so that the ANN- given below (Bishop 1995):
Table 3 MLR coefficients and their significance for the MLR model for site F-6 using a step-wise regression technique (parameters defined
in the text). SE standard error; Temp temperature
Parameter Beta SE (m) B SE (m) t (37) p-level
Intercept 1.625 0.238 6.831 0.000000
Temp 0.259 0.106 0.025 0.010 2.448 0.001921
D4 (Apr–May) 0.236 0.027 0.618 0.070 8.840 0.000000
GWt–1 0.497 0.080 0.495 0.080 6.175 0.000000
D7 (Jul–Aug) 0.173 0.027 0.453 0.071 6.356 0.000000
Rainfall 0.141 0.027 0.001 0.000 5.160 0.000009
Rainfallt –2 −0.074 0.028 0.0003 0.000 −2.621 0.000637
D10 (Oct–Nov) −0.036 0.024 −0.095 0.064 −1.483 0.006509
D2 (Feb–Mar) −0.065 0.025 −0.170 0.066 −2.593 0.001354
Tempt–2 −0.266 0.102 −0.026 0.010 −2.620 0.001267
Tempt–1 0.329 0.157 0.032 0.015 2.089 0.004366
Regression summary: multiple R=0.941, multiple R2 =0.885, adjusted R2 =0.869, F-statistic=216.414, p-level=0.0000 and SEE=0.107 m;
D2, D4 etc. refer to seasonal dummy variables
GWL groundwater-level, R rainfall, T temperature, S river stage, D1, D2, D3,..,D11 seasonal dummy variables
parameter that produced the lowest fitness value during training criteria, such as fixing the number of epochs setting a
runs was selected as the optimal ANN network. target error-goal and fixing the minimum performance
gradient (derivatives of network error with respect to those
Optimization of neuron weights weights and bias). An initial weight range of −0.5 to 0.5 was
During the training process, the three factors that are selected in this study, as suggested by earlier researchers
associated with optimization of weights are: (1) initial (e.g., Sietsma and Dow 1991; Looney 1996). These weights
weight matrix; (2) error function; and (3) termination were then optimized using Eq. (11). The following objective
function (E) was used for the optimization of weights error (RMSE), coefficient of determination (r2) and Nash-
(Principe et al. 2000): Sutcliffe efficiency (NSE), in order to examine their
effectiveness in predicting transient groundwater levels at
1 X 2
n
individual sites. MAE indicates the average of the total
E¼ hoi −hpi ð15Þ model errors (both under-prediction and over-prediction)
2n i¼1
and is used to measure how close forecasts/predictions are to
where, hoi =observed groundwater-level at the ith time [L], the eventual outcomes (observed values). RMSE indicates
hpi =predicted groundwater-level at the ith time [L], and an overall (global) discrepancy between the observed values
n=total number of observations. and the calculated (predicted or simulated) values. r2
Monthly groundwater-level data for the 4-year period describes the proportion of the total variance in the observed
1999–2002 were used for training the 17 site-specific ANN data that can be explained by a model. NSE, on the other
models, with the 2-year period 2003–2004 being used for hand, is an index for assessing the predictive accuracy of a
testing. The entire ANN modeling exercise was carried out model, and it represents an improvement over r2 for model
using the Neural Builder Wizard of NeuroSolutions (version evaluation, because it is sensitive to the differences in
5.05) software (Principe et al. 2000). observed and model-simulated or predicted means and
variances (Legates and McCabe 1999). The mathematical
expressions for MAE, RMSE, r2 and NSE are as follows:
1X
n
Model evaluation criteria
Performance evaluation of the 17 site-specific MLR and MAE ¼ hoi −hpi ð16Þ
n i¼1
ANN models was carried out using four statistical indicators,
namely: mean absolute error (MAE), root mean squared
Table 8 Goodness-of-fit statistics for the ANN and MLR models of the 17 sites during training and testing periods
Site Dataset used ANN models MLR models
MAE (m) RMSE (m) r2 NSE MAE (m) RMSE (m) r2 NSE
A-2 Training 0.072 0.05 0.987 0.984 0.122 0.078 0.963 0.963
Testing 0.121 0.128 0.935 0.912 0.137 0.127 0.914 0.914
B-3 Training 0.023 0.033 0.925 0.907 0.041 0.052 0.780 0.780
Testing 0.038 0.04 0.825 0.822 0.056 0.050 0.736 0.736
C-2 Training 0.075 0.085 0.999 0.995 0.213 0.304 0.947 0.947
Testing 0.311 0.397 0.918 0.901 0.297 0.474 0.858 0.858
C-7 Training 0.021 0.036 0.832 0.828 0.057 0.059 0.544 0.544
Testing 0.042 0.038 0.781 0.776 0.061 0.080 0.007 0.007
DOI 10.1007/s10040-013-1029-5
1877
Fig. 7 Observed and predicted groundwater levels using ANN and MLR models at site B-3: a for the training period (1999–2002); b for
the testing period (2003–2004) (m MSL m above mean sea level)
of known variance to each sample, while keeping the where, Sk =sensitivity index for input k, yip ¼ ith
other inputs at their mean values and then calculating the output obtained with the fixed weights for the pth
change in the model output (i.e., simulated water level). pattern, n=number of network outputs, p=number of
The changes in the model results were assessed using a patterns, and σk2 = variance of the input k. The
sensitivity index, which is expressed as follows (Principe training dataset was used for computing sensitivity
et al. 2000): indices of each input. After calculating the sensitivity
Xp X n 2 index for all the inputs at each of the 17 sites, they
yip −yip were classified into five categories, depending upon
p¼1 i¼1 their sensitivity index values and ranked on a scale of
Sk ¼ ð20Þ
σk 2 1–5 (Table 2).
Fig. 8 Observed and predicted groundwater levels using ANN and MLR models at site E-4: a for the training period (1999–2002); b for
the testing period (2003–2004)
Fig. 9 Observed and predicted groundwater levels using ANN and MLR models at site GH-4.5: a for the training period (1999–2002); b
for the testing period (2003–2004)
Fig. 10 Scatter plots of observed and predicted groundwater levels Fig. 11 Scatter plots of observed and predicted groundwater levels
by ANN and MLR models with ± 1 % error band at site B-3: a for by ANN and MLR models with ± 10 % error band at site E-4: a for
the training period (1999–2002); b for the testing period (2003– the training period (1999–2002); b for the testing period (2003–
2004) (r correlation coefficient between observed and simulated 2004) (r correlation coefficient between observed and simulated
groundwater-level elevation) groundwater-level elevation)
Fig. 13 Comparison of prediction errors of groundwater levels for ANN and MLR models at: a site B-3; b site E-4; c site GH-4.5; all for
the testing period (2003–2004) (Linear line represents best linear fit trendline for ANN and MLR models)
Comparative performance of the ANN and MLR models corresponding MLR models during both training and testing
periods by using MAE, RMSE, r2 and NSE goodness-of-fit
Quantitative evaluation statistics. The results of this comparison are presented in Table 8.
The performances of the 17 site-specific ANN models in Like the training results, the testing results also indicated
predicting groundwater levels were compared with those of the appreciably high values of r2 and NSE and considerably low
Rank 1: very high sensitivity (Sk >0.1); Rank 2: high sensitivity (Sk =>0.05−0.1); Rank 3: moderate sensitivity (Sk =>0.01–0.05); Rank 4: low sensitivity (Sk =>0.005–0.01); Rank 5: very
low sensitivity (Sk ≤0.005). The rank 5 values are italicized
D1–D11 indicate seasonal dummy variables
DOI 10.1007/s10040-013-1029-5
1883
1884
Fig. 14 Sensitivity of the MLP-LM model to the input variables for: a site E-4; b site G-2. In these graphs, D1–D11 indicate seasonal
dummy variables
values of MAE and RMSE for the ANN models compared to NSE=0.998) and site H-2 (MAE=0.017 m, RMSE=0.02 m,
those for the MLR models. The ANN (MLP-LM) models r2 =0.997, NSE=0.997), whereas the prediction accuracy at
have the highest overall prediction accuracy at site GH-4.5 site C-7 is the lowest (MAE=0.021 m, RMSE=0.036 m,
(MAE = 0.015 m, RMSE = 0.03 m, r 2 = 0.998 and r2 =0.832 and NSE=0.828), but it is still acceptable,