Next Article in Journal
Cutting Force and Cutting Quality during Tapered Milling of Glass Magnesium Board
Next Article in Special Issue
Preparation and Removal Properties of Cimetidine from Aqueous Solution by Waste Bricks Incorporated with Different Iron Oxides
Previous Article in Journal
A Structure for Accurately Determining the Mass and Center of Gravity of Rigid Bodies
Previous Article in Special Issue
The Influence of Temperature on Metabolisms of Phosphorus Accumulating Organisms in Biological Wastewater Treatment Plants in the Presence of Cu(II) Toxicity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Can Decomposition Approaches Always Enhance Soft Computing Models? Predicting the Dissolved Oxygen Concentration in the St. Johns River, Florida

1
Department of Water Engineering, Shahid Bahonar University of Kerman, Kerman 76169-14111, Iran
2
Department of Constructional and Environmental Engineering, Kyungpook National University, Sangju 37224, Korea
3
Department of Railroad Construction and Safety Engineering, Dongyang University, Yeongju 36040, Korea
4
Department of Water Engineering, University of Tabriz, Tabriz 51666-16471, Iran
5
Department of Land, Water and Environment Research, Korea Institute of Civil Engineering and Building Technology, Goyang 10223, Korea
6
Department of Biological and Agricultural Engineering & Zachry Department of Civil Engineering, Texas A & M University, College Station, TX 77843-2117, USA
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(12), 2534; https://doi.org/10.3390/app9122534
Submission received: 19 May 2019 / Revised: 11 June 2019 / Accepted: 15 June 2019 / Published: 21 June 2019

Abstract

:
This study evaluates standalone and hybrid soft computing models for predicting dissolved oxygen (DO) concentration by utilizing different water quality parameters. In the first stage, two standalone soft computing models, including multilayer perceptron (MLP) neural network and cascade correlation neural network (CCNN), were proposed for estimating the DO concentration in the St. Johns River, Florida, USA. The DO concentration and water quality parameters (e.g., chloride (Cl), nitrogen oxides (NOx), total dissolved solid (TDS), potential of hydrogen (pH), and water temperature (WT)) were used for developing the standalone models by defining six combinations of input parameters. Results were evaluated using five performance criteria metrics. Overall results revealed that the CCNN model with input combination III (CCNN-III) provided the most accurate predictions of DO concentration values (root mean square error (RMSE) = 1.261 mg/L, Nash-Sutcliffe coefficient (NSE) = 0.736, Willmott’s index of agreement (WI) = 0.919, R2 = 0.801, and mean absolute error (MAE) = 0.989 mg/L) for the standalone model category. In the second stage, two decomposition approaches, including discrete wavelet transform (DWT) and variational mode decomposition (VMD), were employed to improve the accuracy of DO concentration using the MLP and CCNN models with input combination III (e.g., DWT-MLP-III, DWT-CCNN-III, VMD-MLP-III, and VMD-CCNN-III). From the results, the DWT-MLP-III and VMD-MLP-III models provided better accuracy than the standalone models (e.g., MLP-III and CCNN-III). Comparison of the best hybrid soft computing models showed that the VMD-MLP-III model with 4 intrinsic mode functions (IMFs) and 10 quadratic penalty factor (VMD-MLP-III (K = 4 and α = 10)) model yielded slightly better performance than the DWT-MLP-III with Daubechies-6 (D6) and Symmlet-6 (S6) (DWT-MLP-III (D6 and S6)) models. Unfortunately, the DWT-CCNN-III and VMD-CCNN-III models did not improve the performance of the CCNN-III model. It was found that the CCNN-III model cannot be used to apply the hybrid soft computing modeling for prediction of the DO concentration. Graphical comparisons (e.g., Taylor diagram and violin plot) were also utilized to examine the similarity between the observed and predicted DO concentration values. The DWT-MLP-III and VMD-MLP-III models can be an alternative tool for accurate prediction of the DO concentration values.

1. Introduction

Water quality explains the descriptions of biological, chemical, and physical characteristics of water bodies [1,2]. The assessment of water quality parameters, such as dissolved oxygen (DO), algae, nitrogen (N), total nitrogen (TN), phosphorus (P), total phosphorus (TP), biochemical oxygen demand (BOD), and chemical oxygen demand (COD), is necessary to improve the operational performance and develop water resource management effectively [3].
DO, as one of the water quality parameters, refers to the free-level non-compound oxygen dissolved in water or other liquids [4,5,6,7,8]. Too high or low levels of DO concentration can affect the maintenance of water quality [9]. Therefore, DO is a key parameter to assess water quality in rivers, reservoirs, and lakes, and as one of the indicators for the healthy functioning of aquatic ecosystems [10].
Water quality modeling includes contaminant transport, biochemical transformation, and forecasting/prediction of water pollution [3,11]. Accurate forecasting/prediction of water quality can provide the basic data to control water quality and deal with water quality incidents [12]. In traditional water quality modeling, water quality models can be calibrated using the trial and error method. However, since the traditional calibration processes require a large number of iterations, they cannot be effective [11]. Therefore, the inverse methods (e.g., fuzzy logic, Bayesian inference, and maximum entropy technique, etc.), which are more robust, objective, and sound approaches, are recommended. One of the widely used inverse methods is the indirect inverse method, which is developed using forging nonlinear optimization problems [11,13].
The diverse researches on water quality modeling using heuristic approaches (e.g., artificial neural network (ANN), adaptive neuro-fuzzy inference systems (ANFIS), and support vector machines (SVM), etc.) have been investigated and reported in the literature during the past two decades [14,15,16,17,18,19,20].
Diamantopoulou et al. [21] suggested the cascade correlation neural network (CCNN) to predict the missing monthly data of water quality parameters. Kuo et al. [22] suggested the backpropagation neural network (BPNN) to quantify the cause-and-effect relationship in reservoir eutrophication. Zhao et al. [23] developed the BPNN to forecast water quality in reservoirs. Zou et al. [11] developed the genetic algorithm-based neural network (GA-NN) to solve inverse water quality problems. Dogan et al. [24] predicted the BOD concentration using the feedforward neural network (FFNN) in the Melen River, Turkey. Najah et al. [3] used ANNs to predict three water quality parameters (i.e., total dissolved solids, electrical conductivity, and turbidity) in the Johor River, Malaysia.
Singh et al. [25] employed SVM (i.e., support vector classification (SVC) and regression (SVR)) to optimize a monitoring program using water quality data. Han et al. [26] used a radial basis function neural network (RBFNN) for water quality prediction in the wastewater treatment process. Gazzaz et al. [9] proposed the FFNN to predict water quality indexes in the Kinta River, Malaysia. Xu and Lie [12] provided a hybrid approach combining the BPNN and wavelet transform (WT) to develop a water quality model. Ay and Kisi [27] developed the k-means clustering-based multilayer perceptron (k-means MLP) for modeling the COD concentration in Adapazari, Turkey. Li et al. [28] implemented a hybrid model based on integrated SVR with the firefly algorithm (FFA) for prediction of the water quality indicator (WQI) over a period of 10 years in the Euphrates River, Iraq.
Among the various heuristic techniques, ANNs (e.g., BPNN, MLP, and FFNN) have been accomplished for prediction, forecasting, modelling, and estimation of DO [29,30]; BOD and DO [31]; TN, TP, and DO [32] using different input data in the river stream. Some heuristic approaches (e.g., MLP, FFNN, ANFIS, and RBFNN) have been applied to ponds, lakes, and reservoirs [5,33,34,35,36,37].
Singh et al. [4] introduced two ANNs for estimating water quality parameters (i.e., DO and BOD) in the Gomti River, India. Faruk [38] proposed an autoregressive integrated moving average (ARIMA) based neural network (ARIMA-NN) to predict water quality parameters (e.g., temperature, boron, and DO) in the Büyük Menderes River, Turkey. Ay and Kisi [10] proposed the MLP, RBFNN, and multilinear regression (MLR) to estimate the DO concentration in Fountain Creek, Colorado. Han et al. [39] developed a self-organizing RBF (SORBF) to predict the DO concentration in activated sludge wastewater treatment processes. Martí et al. [40] proposed the MLP, gene expression programming (GEP), and MLR to estimate the DO concentration at a sand filter outlet. Antanasijević et al. [41] used the generalized regression neural network (GRNN) for forecasting the DO concentration in the Danube River, Serbia. Heddam [42] proposed the grid partition-based ANFIS (ANFIS-GRID), subtractive clustering based ANFIS (ANFIS-SUB), and MLR for estimating the DO concentration in the Klamath River, Oregon. Najah et al. [43] applied the ANFIS for the prediction of DO concentration. Nemati et al. [44] investigated ANFIS, MLP, and MLR to estimate the DO concentration in the Tai Po River, Hong Kong. Keshtegar and Heddam [45] developed the modified response surface method (MRSM) and MLP to estimate the DO concentration.
Although there have been many studies to estimate the DO concentration using heuristic approaches, the applications of CCNN have been limited. Olyaie et al. [8] used the MLP, RBFNN, SVM, and linear genetic programming (LGP) for predicting the DO concentration in the Delaware River, New Jersey. Tomić et al. [46] determined the extrapolation of an ANN model, which was established for predicting the DO concentration in the Danube River. In addition, a specific investigation using heuristic and decomposition approaches, for example, the hybrid approaches using DWT [47,48,49,50] and VMD [51], cannot be found with ease for DO forecasting/prediction category [12,52].
Liu et al. [48] proposed a hybrid model based on wavelet analysis (WA), least squares support vector regression (LSSVR), and an optimal improved Cauchy particle swarm optimization (CPSO) algorithm for predicting the DO concentration. The WA-LSSVR-CPSO model demonstrated a powerful and reliable approach for predicting the DO concentration in intensive aquaculture. Ravansalar et al. [50] evaluated 30-min DO concentration using the ANN and wavelet-based ANN (WANN) models in the River Calder, England. Results showed that the WANN model provided a 30-min DO concentration prediction that was comparable to ANN. Fijani et al. [51] discussed two hybrid models based on the complete ensemble empirical mode decomposition algorithm with adaptive noise (CEEMDAN), VMD, extreme learning machines (ELMs), and least squares support vector machines (LSSVMs) models for predicting chlorophyll-a (Chl-a) and DO concentrations, respectively. They found that the CEEMDAN-VMD-ELM model produced the best results to predict water quality parameters (e.g., Chl-a and DO concentration).
This paper proposes two heuristic models (e.g., MLP and CCNN) and decomposition approaches (e.g., DWT and VMD) for predicting the DO concentration in the St. Johns River, Florida, USA. The model performances are evaluated using model efficiency indexes and diagnostic analysis using graphical comparisons (e.g., Taylor diagram and violin plot). The paper is organized as follows: The second part provides methodologies, including MLP, CCNN, DWT, and VMD, respectively. The third part proposes a study area and data, and the fourth part presents results and discussion. Conclusions are found at the end of the paper.

2. Materials and Methods

2.1. Multilayer Perceptron (MLP) Neural Network Model

ANNs have been adopted as popular and well-established models in heuristic approaches. An ANN is a parallel information processing system with a set of neurons arranged in one or more hidden layers [53]. The MLP, which is an explicit form of the ANNs model, consists of three (or more) layers with an input layer (where the data are fed into the model), one (or more) hidden layer (where the data are processed to construct a model), and an output layer (where the results are generated) [54,55,56,57]. The neurons are connected by appropriate weights in each layer to the neurons in continuous layers. In this study, the sigmoid and linear activation functions, which have been commonly utilized for modeling purposes [58,59,60,61], were utilized in the hidden and output layers, respectively. Moreover, the Levenberg–Marquardt backpropagation (LMBP) algorithm was employed for training the MLP model. The LMBP algorithm has a more accurate curve fitting ability that is applied to the input–output data [62,63]. Figure 1a shows the structure of the MLP model, where, i, j, k = the input layer, the hidden layer, and the output layer, respectively; Wkj = the connection weights between the hidden and the output layers; Wji = the connection weights between the input and the hidden layers; B1 = the bias in the hidden layer; and B2 = the bias in the output layer. In addition, DO = dissolved oxygen (mg/L); Cl = chloride (mg/L); NOx = nitrogen oxides (mg/L); TDS = total dissolved solid; pH = potential of hydrogen; and WT = water temperature (°C).

2.2. Cascade Correlation Neural Network (CCNN) Model

A CCNN is an efficient constructive neural network combining the idea of incremental structure and learning during its training. Training starts with a minimal network consisting of an input and output layer without a hidden layer. If the training can no longer reduce the residual error, then the training phase is stopped, and enters the next phase for the training of a potential hidden neuron [64,65]. The potential hidden neuron has associated connection weights from the input layer and all preexisting hidden neurons but not toward the output layer. The connection weights are optimized by the gradient ascent method to maximize the correlation between its output and the residual error of the CCNN model. When a potential hidden neuron is trained, connection weights associated with the output layer remain unchanged. When a potential hidden neuron is added to the structure of the CCNN model, it becomes a new hidden neuron, and its incoming connection weights are fixed for the remainder of the training phase [65,66,67]. Figure 1b represents the structure of the CCNN model.

2.3. Discrete Wavelet Transform (DWT)

Wavelet transform decomposition (WTD) can be generally classified as continuous wavelet transform (CWT) and discrete wavelet transform (DWT) [68,69]. DWT requires less time of the arithmetic processes, and is easier to implement than CWT [68,70]. A fast DWT algorithm requires four filters for perfect implementation (e.g., decomposition low-pass, decomposition high-pass, reconstruction low-pass, and reconstruction high-pass) [68,71,72,73]. The low-pass filter for decomposition and reconstruction categories permits the interpretation of low frequency components, while the high-pass filter approves the investigation of high frequency components [72,74]. The multi-resolution approach using Mallat’s DWT algorithm can be explained as a process to depict ‘approximation’ and ‘details’ for an underlying signal. An approximation produces a conventional trend of the original signal, while the details provide its high-frequency components [72,73,75]. The feature reports for Mallat’s DWT algorithm can be found in Nason [76] and Percival and Walden [77]. Figure 2 shows Mallat’s DWT algorithm for three-level decomposition [73].

2.4. Variational Mode Decomposition (VMD)

VMD is a fully adaptive and non-recursive algorithm for time-frequency signal analysis [78]. An original time series, f, can be decomposed into K intrinsic mode functions (IMFs) using the VMD approach. The constrained variation formulation for generating IMFs can be written as Equation (1):
min { u k } , { ω k } { k = 1 K t [ ( δ ( t ) + j π t ) u k ( t ) ] e j ω k t 2 2 } , s . t . k = 1 K u k ( t ) = f ,
where δ = the Dirac function; j 2 = 1 ; 2 = the L 2 distance; ω k = the center frequency; * = the convolution; u k ( t ) = A k ( t ) cos ( ϕ k ( t ) ) = the kth IMF; ϕ k = the non-decreasing function; and A k = the non-negative function. The constrained variational formulation can be modified as the following unconstrained pattern using an augmented Lagrangian method [78,79]:
L ( { u k } , { ω k } , λ ) = α k = 1 K t [ ( δ ( t ) + j π t ) u k ( t ) ] e j ω k t 2 2 + f ( t ) k = 1 K u k ( t ) 2 2 + λ ( t ) , f ( t ) k = 1 K u k ( t )
where L = the augmented Lagrangian; λ = the Lagrange multiplier; and a , b = the scalar product of a and b. Figure 3 explains the flowchart of the VMD algorithm [73].

2.5. Hybrid Modeling Using DWT and VMD Approaches

DWT-based soft computing models (DWT-MLP and DWT-CCNN) are the hybrid models combined with the standalone models (MLP and CCNN) and DWT, respectively. In the same manner, VMD-based soft computing models (VMD-MLP and VMD-CCNN) conjugate the standalone models (MLP and CCNN) and VMD, respectively. Therefore, DWT- and VMD-based soft computing models consist of three steps (1). The training and testing dataset are decomposed into an approximation and multiple details using the DWT approach, and multiple IMFs using VMD, respectively (2). The standalone models (MLP and CCNN) are developed for each decomposed training dataset (3). The final predictions of DO concentration values are obtained by aggregating the sub-time series predicted from the standalone models (MLP and CCNN), respectively. Figure 4 represents the flowchart for DWT- and VMD-based soft computing modeling.

2.6. Performance Evaluation of Models

Performance measures are assessed by comparing predicted values with their corresponding observed values using the following criteria:
  • Root mean square error (RMSE):
    RMSE = 1 n i = 1 n [ D O o b s D O p r e ] 2 .
  • Nash–Sutcliffe coefficient (NSE):
    NSE = 1 i = 1 n [ D O o b s D O p r e ] 2 i = 1 n [ D O o b s D O ¯ p r e ] 2 .
  • Willmott’s index of agreement (WI):
    WI = 1 [ i = 1 n ( D O o b s D O p r e ) 2 i = 1 n ( | D O p r e D O ¯ o b s | + | D O o b s D O ¯ o b s | ) 2 ] .
  • Mean absolute error (MAE):
    MAE = 1 n i = 1 n | D O p r e D O o b s | .
  • Coefficient of determination (R2):
    R 2 = ( [ i = 1 n D O o b s D O p r e 1 n i = 1 n D O o b s i = 1 n D O p r e ] [ i = 1 n D O o b s 2 1 n [ i = 1 n D O o b s ] 2 ] [ i = 1 n D O p r e 2 1 n [ i = 1 n D O p r e ] 2 ] ) ,
    where D O o b s and D O p r e are the observed and predicted values, respectively; D O ¯ o b s and D O ¯ p r e are the average of observed and predicted values, respectively; and n is the length of time series data. The discrepancy between observed and predicted values can be shown using the RMSE criterion. A value of zero reflects perfect prediction. The RMSE criterion must be used for model evaluation to obtain accuracy in absolute units [80]. NSE is taken into account to evaluate the ability of predicting models [81]. If the squared difference between observed and predicted DO values is relatively large to concur with the variance in the observed DO values, then the NSE criterion will be zero. If the NSE criterion is negative, the results indicate that the observed mean is a better predictor than the model [81,82]. If the NSE criterion is equal to one, it indicates a perfect model [83]. The WI criterion varies between zero and one. WI calculates the ratio of mean square error (MSE) and can provide an advantage over the RMSE [84,85,86]. The MAE criterion can provide better information for a model’s prediction. The MAE cannot be weighted towards higher or lower magnitudes. However, it evaluates all derivations from observed DO values in an equal manner [87].

3. Case Study

In this study, the fluctuations of the DO concentration independent of some parameters (e.g., chloride (Cl), nitrogen oxides (NOx), total dissolved solid (TDS), potential of hydrogen (pH), and water temperature (WT)) were selected as a case study. This study area was located in the southern east of Florida with a latitude of 28°32′33.864″ N and a longitude of 80°56′33.428″ W (Figure 5). All data numbered 232 records along about 12 years (1996–2013). The data were arbitrarily divided into two parts for training and testing phases. The training datasets were chosen at 80% (n = 186) of the data length and the testing datasets covered the remaining 20% (n = 46).

4. Application and Results

4.1. Setting up the Standalone Models

The statistical parameters of the collected dataset are presented in Table 1. A considerable wide domain of data values can be observed in Table 1 (e.g., max of NOx = 0.46 mg/L and max of TDS = 1950.0). This issue implies that before importing the data to the standalone models (MLP and CCNN), they should be standardized between specific ranges (e.g., from 0 to 1). As can be seen from Table 1, the pH was the highest correlated parameter to the DO concentration (correlation coefficient (CC) = 0.760). It was followed by NOx (CC = 0.554) and then by WT with a negative value of the correlation coefficient (CC = −0.544), which denoted the reverse effect of WT on the DO concentration. Based on the coefficient of variation, the WT data were least dispersed, and the NOx data were most sporadic. Several combinations for setting up the standalone models (MLP and CCNN) regarding the input vectors (e.g., consisting of five chemical characteristic parameters of chloride (Cl), nitrogen oxides (NOx), total dissolved solid (TDS), potential of hydrogen (pH), and water temperature (WT)) can be created. Bear in mind that the only value in the output layer always corresponded to the DO concentration.
In this study, six different input combinations for constructing the models, including five chemical characteristic parameters (combination ID = I) and even one quality parameter (combination ID = V and combination ID = VI), based on the highest positive and negative values of CC in Table 1 were constructed (see Table 2).
Constructing the ANN architecture involves the creation of the ANN topology and training parameters, such as the number of neurons in the hidden layer(s). Looking back at Table 2, the first step of ANN architecture for determining the input combinations was already completed. In the second step, the main structure of ANN in terms of the number of layers should be specified. Based on the reports for the capability of the one hidden layer supervised neural networks in simulating complex phenomena [88], this study adopted a one hidden layer architecture for the standalone models (e.g., MLP and CCNN). Finally, the optimal number of neurons in the hidden layer was determined using the MSE criterion by a trial and error approach (see the third column of Table 3, Table 4 and Table 5).

4.2. Performance of Standalone Models

A statistical summary of the DO concentration performance using the standalone models (MLP and CCNN) is given in Table 3. Results of the MLP model with higher (NSE, WI, and R2) and lower (RMSE and MAE) values indicated that the first (MLP-I) and third (MLP-III) input combinations acted better than the others for the training and testing phases, respectively. Since, however, choosing the best model is always based on the performance of the testing phase, the MLP model with the third combination topology (MLP-III) was selected as the best model. Similar interpretation can be done for the CCNN models. While the second combination (CCNN-II) provided the best performance for the training phase, the third input combination (CCNN-III) gave the best results for the testing phase. All the statistics were enriched in the CCNN-III model in comparison to the MLP-III model. A general comparison of the MLP-III and CCNN-III models revealed that the CCNN-III model yielded the better predictions than the MLP-III model. The RMSE, NSE, WI, R2, and MAE criteria for the testing phase of the CCNN-III model were improved by 129%, 8%, 2%, 3%, and 434% compared with the MLP-III model, respectively.
The variations of the observed versus predicted DO concentration values are depicted in Figure 6a,b. Visual analysis confirmed the better results of the CCNN-III model compared to those of the MLP-III model with he observed values. Figure 7a,b provides scatter plots of the testing dataset between the observed and predicted values using the MLP-III and CCNN-III models. From visual interpretation of Figure 7a,b, it can be understood that the dots in the MLP-III plot were a bit more sporadic than those in the CCNN-III plot. In addition, based on the coefficient of determination and the slope of the trend lines to the unity, it can be concluded that the CCNN-III model acted better in predicting the DO concentration.
Figure 8 displays the error residual in mg/L over the testing phases of the MLP-III and CCNN-III modeling. It explained that the residuals of local peaks were relatively insignificant for the CCNN-III model, and tended to overestimate the DO concentration values; whereas the MLP-III model had larger residuals, and tended to underestimate the DO concentration values.

4.3. Performance of Hybrid Models

4.3.1. DWT-Based Soft Computing Models

To decompose the input dataset using the DWT algorithm, the optimal level of decomposition (L) should be selected. In this study, Equation (8) was used to calculate the optimal level of decomposition [68,70,72]. Although the optimal level of decomposition can be implemented using a trial-and-error method, it is time-consuming and a waste of energy:
L = int [ log ( N ) ] ,
where N is the length of the time series, int [ k ] returns the integer portion of k, and k is a real number.
In this study, L = 2 was determined using Equation (8). In addition, mother wavelets have to be set before DWT-based soft computing models are employed. Using different mother wavelets, the dataset was decomposed with a details (D1 and D2) and an approximation (A2) components for individual input data [72,75]. For the DWT algorithm, Daubechies, Symmlets, and Coiflets have been frequently used as mother wavelets in previous studies [72,75,89,90]. Therefore, the underlying mother wavelets, including Coiflet-6 (C6), Coiflet-12 (C12), Coiflet-18 (C18), Daubechies-6 (D6), Daubechies-12 (D12), Daubechies-18 (D18), Symmlet-6 (S6), Symmlet-12 (S12), and Symmlet-18 (S18), were implemented. For each DWT-based soft computing model, the optimal mother wavelet yielding the best model performance was recommended. Figure 9 shows an approximation and details decomposed using the Symmlet-6 (S6) mother wavelet for the original water temperature (WT).
Table 4 gives the performance statistics using the DWT-MLP-III and DWT-CCNN-III models during the training and testing phases. Table 3 and Table 4 suggested that all of the DWT-MLP-III models improved the performance of the MLP-III model significantly, while all of the DWT-CCNN-III models did not improve the performance of the CCNN-III model during the testing phase. In addition, the DWT-MLP-III (D6) and DWT-MLP-III (S6) (e.g., RMSE = 0.161 (mg/L), NSE = 0.983, WI = 0.996, R2 = 0.983, and MAE = 0.061 (mg/L) for D6 and S6) models produced the best results among all of the DWT-MLP-III models during the testing phase. The combination of the DWT into the MLP-III model could confirm the model accuracy for the prediction of the DO concentration. However, the CCNN-III model provided better results than any of the DWT-CCNN-III models. A comparison explained that all the DWT-MLP-III models yielded better results compared with all the DWT-CCNN-III models.
Figure 10a,b shows the scatter plots of the testing dataset between the observed and predicted values using the DWT-MLP-III (S6) and DWT-CCNN-III (S6) models. From visual interpretation of Figure 10a,b, it can be explained that the dots in the DWT-CCNN-III (S6) plot were extremely sporadic compared to those in the DWT-MLP-III (S6) plot. In addition, based on the coefficient of determination and the slope of trend lines to the unity, it can be concluded that the DWT-MLP-III (S6) model acted better in predicting the DO concentration.

4.3.2. VMD-Based Soft Computing Models

To decompose the input dataset using the VMD algorithm, the number of IMFs (K) and the quadratic penalty factor (α) have to be implemented in advance. In this study, different sets of parameters were investigated, and three sets of parameters with higher correlations between the original and predicted data (i.e., the aggregation of decomposed series) were selected. The three sets were (K, α) = {(3, 5), (4, 5), (4, 10)}. Among them, the optimal parameters (K = 4 and α = 10) yielding the best performance of the VMD-based soft computing models were chosen finally. Figure 11 shows the original WT series and the IMFs decomposed using the VMD algorithm (K = 4 and α = 10).
Table 5 shows the performance statistics using the VMD-MLP-III and VMD-CCNN-III models during the training and testing phases. Table 3 and Table 5 indicate that all of the VMD-MLP-III models enhanced the performance of the MLP-III model enormously, while all of the VMD-CCNN-III models did not enhance the performance of the CCNN-III model during the testing phase. In addition, the VMD-MLP-III (K = 4 and α = 10) model (e.g., RMSE = 0.107 (mg/L), NSE = 0.992, WI = 0.998, R2 = 0.993, and MAE = 0.034 (mg/L)) provided the best results among all of the VMD-MLP-III models during the testing phase. The conjugation of VMD into the MLP-III model ensured the model accuracy for predicting the DO concentration. However, the CCNN-III model provided better results compared to all of the VMD-CCNN-III models. A comparison explained that all of the VMD-MLP-III produced better results compared with all of the VMD-CCNN-III models.
Table 4 and Table 5 explain that the statistical results of the DWT- and VMD-MLP-III models showed similar statistical patterns. A comparison of the best models revealed, however, that the VMD-MLP-III (K = 4 and α = 10) model yielded slightly better results than the DWT-MLP-III (D6 and S6) model. Figure 12a,b shows scatter plots of the testing dataset between the observed and predicted values using the VMD-MLP-III (K = 4 and α = 10) and VMD-CCNN-III (K = 4 and α = 10) models. From visual interpretation of Figure 12a,b, it can be explained that the dots in the VMD-CCNN-III (K = 4 and α = 10) plot were extremely sporadic compared to those in the VMD-MLP-III (K = 4 and α = 10) plot. Based on the coefficient of determination and the slope of trend lines to the unity, it can be concluded that the VMD-MLP-III (K = 4 and α = 10) model acted better in predicting the DO concentration.
In general, various studies have reported that the combination of soft computing models and decomposition approaches improved the accuracy and reliability of model performance for predicting DO concentration [12,47,48,49,50,51,52]. Even if the CCNN model showed the outstanding performance for standalone models, the combination of the CCNN model and decomposition approaches cannot improve the model performance. The special model structure (e.g., adding hidden nodes) can prevent the model performance of complex nonlinear signals. To confirm the model performance, continued studies are required using different data, soft computing models, and decomposition approaches for predicting diverse environmental parameters (e.g., BOD, COD, TP, and TN etc.).

4.4. Diagnostic Analysis

In this study, three diagnostic analysis methods (i.e., Taylor diagram and violin plot) were considered for visual evaluation of the model performance.

4.4.1. Taylor Diagram

A polar plot presented by Taylor [91] was drawn for obtaining a visual understanding of model performance. It has the ability to highlight the goodness of model performance in comparison to observed values. The Taylor diagram depicts three statistics: (1) Correlation coefficient (the azimuth angle), (2) normalized standard deviation (radial distance from the origin), and (3) RMSE (distance from the reference observed point). A perfect matching of the predicted results is identified as a complete overlay by the reference point with the correlation coefficient equal to unity and the exact amplitude of variations compared with observations [91,92,93,94].
Figure 13a shows the Taylor diagram of the standalone models (MLP and CCNN). In the case of the best models, the diagram shows that the CCNN-III model had a lower RMSE than the MLP-III model. Although the correlation coefficients and standard deviations of the predicted data for both models were less than the observations, the node representing the CCNN-III model was closer to the observation node. Figure 13b shows the Taylor diagram of the hybrid models. The diagram shows that the DWT-MLP-III (S6) and VMD-MLP-III (K = 4 and α = 10) models had a lower RMSE than the DWT-CCNN-III (S6) and VMD-CCNN-III (K = 4 and α = 10) models. Although the correlation coefficients and standard deviations of the predicted data of the DWT-MLP-III (S6) and VMD-MLP-III (K = 4 and α = 10) models were less than observations, the node representing the VMD-MLP-III (K = 4 and α = 10) model was closer to the observation node.

4.4.2. Violin Plot

As a further diagnostic tool, the violin plot is utilized in Figure 14a,b to assess the predicted results of the developed models for DO concentration. The violin plot, which has the ability to indicate the probability distribution of an observed and predicted dataset, is categorized as a box plot with the integration of the kernel density plot [95]. Based on the legends of Figure 14a, the median of the observed data was predicted by the MLP-III accurately (6.173 vs. 6.403), while the 25th and 75th percentiles in the CCNN-III had a better fit than the MLP-III. In addition, the MLP-III model overestimated the minimum, 25th percentile, median, and 75th percentile range of the DO concentration, whereas the CCNN-III model underestimated the 25th percentile, median, and 75th percentile range of the DO concentration. Overall, the violin plots indicated that the CCNN-III model performed better than the MLP-III model. Figure 14b explains that the VMD-MLP-III (K = 4 and α = 10) model overestimated the minimum, 25th, and median of the DO concentration, while the DWT-MLP-III (S6) underestimated the minimum, median, 75th percentile, and maximum of the DO concentration. Overall, the violin plots indicated that the VMD-MLP-III (K = 4 and α = 10) model performed slightly better than the DWT-MLP-III (S6) model.

5. Conclusions

This study investigated the accuracy of two heuristic (MLP and CCNN) and decomposition (DWT and VMD) approaches for predicting dissolved oxygen (DO) concentration. To achieve this goal, the DO concentration and five chemical input parameters (Cl, NOx, TDS, pH, and WT) in the St. Johns River, Florida, USA, were used. For training and testing the developed models, the total dataset was divided into 80% and 20%, respectively. Several statistical indices (e.g., RMSE, NSE, WI, R2, and MAE) and diagnostic analyses (e.g., Taylor diagram and violin plot) were used to compare the developed models.
In the first stage, it was found that the CCNN-III (RMSE = 0.550 mg/L, NSE = 0.797, WI = 0.942, R2 = 0.825, and MAE = 0.185 mg/L) and MLP-III (RMSE = 1.261 mg/L, NSE = 0.736, WI = 0.919, R2 = 0.801, and MAE = 0.989 mg/L) provided the best results based on the standalone model category. In addition, a comparison suggested that the CCNN-III model performed better than the MLP-III model. In the second stage, however, it was found that the DWT-MLP-III (D6 and S6) (RMSE = 0.161 mg/L, NSE = 0.983, WI = 0.996, R2 = 0.983, and MAE = 0.061 mg/L) and VMD-MLP-III (K = 4 and α = 10) (RMSE = 0.107 mg/L, NSE = 0.992, WI = 0.998, R2 = 0.993, and MAE = 0.034 mg/L) produced the best results based on the hybrid model category. Unfortunately, the CCNN-III did not improve the performance using the DWT and VMD approaches.

Author Contributions

M.Z.-K. conceived and designed the research; Y.S. implemented DWT and VMD; S.K. applied the CCNN model and wrote the original manuscripts; M.A.G. implemented the MLP model and performed analyses; S.S. and S.N. collected and analyzed the research data; N.W.K. analyzed the results; V.P.S. reviewed and edited the manuscripts.

Funding

A grant from a Strategic Research Project (20190153-001) funded by the Korea Institute of Civil Engineering and Building Technology.

Acknowledgments

The authors would like to appreciate that this research was supported by a grant from a Strategic Research Project (20190153-001) funded by the Korea Institute of Civil Engineering and Building Technology.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liou, S.M.; Lo, S.L.; Wang, S.H. A generalized water quality index for Taiwan. Environ. Monit. Assess. 2004, 96, 32–35. [Google Scholar] [CrossRef]
  2. Khalil, B.; Ouarda, T.B.M.J.; St-Hilaire, A. Estimation of water quality characteristics at ungauged sites using artificial neural networks and canonical correlation analysis. J. Hydrol. 2011, 405, 277–287. [Google Scholar] [CrossRef]
  3. Najah, A.; Elshafie, A.; Karim, O.A.; Jaffar, O. Prediction of Johor River water quality parameters using artificial neural networks. Eur. J. Sci. Res. 2009, 28, 422–435. [Google Scholar]
  4. Singh, K.P.; Basant, A.; Malik, A.; Jain, G. Artificial neural network modeling of the river water quality—A case study. Ecol. Model. 2009, 220, 888–895. [Google Scholar] [CrossRef]
  5. Ranković, V.; Radulović, J.; Radojević, I.; Ostojić, A.; Čomić, L. Neural network modeling of dissolved oxygen in the Gruža reservoir, Serbia. Ecol. Model. 2010, 221, 1239–1244. [Google Scholar] [CrossRef]
  6. Xia, M.; Craig, P.M.; Schaeffer, B.; Stoddard, A.; Liu, Z.; Peng, M.; Zhang, H.; Wallen, C.M.; Bailey, N.; Mandrup-Poulsen, J. Influence of physical forcing on bottom-water dissolved oxygen within Caloosahatchee River Estuary, Florida. J. Environ. Eng. 2010, 136, 1032–1044. [Google Scholar] [CrossRef]
  7. Xia, M.; Jiang, L. Influence of wind and river discharge on the hypoxia in a shallow bay. Ocean Dyn. 2015, 65, 665–678. [Google Scholar] [CrossRef]
  8. Olyaie, E.; Abyaneh, H.Z.; Mehr, A.D. A comparative analysis among computational intelligence techniques for dissolved oxygen prediction in Delaware River. Geosci. Front. 2017, 8, 517–527. [Google Scholar] [CrossRef] [Green Version]
  9. Gazzaz, N.M.; Yusoff, M.K.; Aris, A.Z.; Juahir, H.; Ramli, M.F. Artificial neural network modeling of the water quality index for Kinta River (Malaysia) using water quality variables as predictors. Mar. Pollut. Bull. 2012, 64, 2409–2420. [Google Scholar] [CrossRef]
  10. Ay, M.; Kisi, O. Modeling of dissolved oxygen concentration using different neural network techniques in Foundation Creek, El Paso County, Colorado. J. Environ. Eng. 2011, 138, 654–662. [Google Scholar] [CrossRef]
  11. Zou, R.; Lung, W.S.; Wu, J. An adaptive neural network embedded genetic algorithm approach for inverse water quality modeling. Water Resour. Res. 2007, 43, W08427. [Google Scholar] [CrossRef]
  12. Xu, L.; Liu, S. Study of short-term water quality prediction model based on wavelet neural network. Math. Comput. Model. 2013, 58, 807–813. [Google Scholar] [CrossRef]
  13. Sen, M.K.; Stoffa, P.L. Global Optimization Methods in Geophysical Inversion; Cambridge University Press: Cambridge, UK, 2013. [Google Scholar]
  14. Maier, H.R.; Dandy, G.C. The use of artificial neural networks for the prediction of water quality parameters. Water Resour. Res. 1996, 32, 1013–1022. [Google Scholar] [CrossRef]
  15. Zhang, Q.; Stanley, S.J. Forecasting raw-water quality parameters for the North Saskatchewan River by neural network modeling. Water Res. 1997, 31, 2340–2350. [Google Scholar] [CrossRef]
  16. Aguilera, P.A.; Frenich, A.G.; Torres, J.A.; Castro, H.; Vidal, J.M.; Canton, M. Application of the Kohonen neural network in coastal water management: Methodological development for the assessment and prediction of water quality. Water Res. 2001, 35, 4053–4062. [Google Scholar] [CrossRef]
  17. Zhang, Y.; Pulliainen, J.; Koponen, S.; Hallikainen, M. Application of an empirical neural network to surface water quality estimation in the Gulf of Finland using combined optical data and microwave data. Remote Sens. Environ. 2002, 81, 327–336. [Google Scholar] [CrossRef]
  18. Zou, R.; Lung, W.S.; Guo, H. Neural network embedded Monte Carlo approach for water quality modeling under input information uncertainty. J. Comput. Civ. Eng. 2002, 16, 135–142. [Google Scholar] [CrossRef]
  19. Ha, H.; Stenstrom, M.K. Identification of land use with water quality data in stormwater using a neural network. Water Res. 2003, 37, 4222–4230. [Google Scholar] [CrossRef]
  20. Diamantopoulou, M.J.; Papamichail, D.M.; Antonopoulos, V.Z. The use of a neural network technique for the prediction of water quality parameters. Oper. Res. 2005, 5, 115–125. [Google Scholar] [CrossRef]
  21. Diamantopoulou, M.J.; Antonopoulos, V.Z.; Papamichail, D.M. Cascade correlation artificial neural networks for estimating missing monthly values of water quality parameters in rivers. Water Resour. Manag. 2007, 21, 649–662. [Google Scholar] [CrossRef]
  22. Kuo, J.T.; Hsieh, M.H.; Lung, W.S.; She, N. Using artificial neural network for reservoir eutrophication prediction. Ecol. Model. 2007, 200, 171–177. [Google Scholar] [CrossRef]
  23. Zhao, Y.; Nan, J.; Cui, F.Y.; Guo, L. Water quality forecast through application of BP neural network at Yuqiao reservoir. J. Zhejiang Univ.-Sci. A 2007, 8, 1482–1487. [Google Scholar] [CrossRef]
  24. Dogan, E.; Sengorur, B.; Koklu, R. Modeling biological oxygen demand of the Melen River in Turkey using an artificial neural network technique. J. Environ. Manag. 2009, 90, 1229–1235. [Google Scholar] [CrossRef] [PubMed]
  25. Singh, K.P.; Basant, N.; Gupta, S. Support vector machines in water quality management. Anal. Chim. Acta 2011, 703, 152–162. [Google Scholar] [CrossRef] [PubMed]
  26. Han, H.G.; Chen, Q.L.; Qiao, J.F. An efficient self-organizing RBF neural network for water quality prediction. Neural Netw. 2011, 24, 717–725. [Google Scholar] [CrossRef] [PubMed]
  27. Ay, M.; Kisi, O. Modelling of chemical oxygen demand by using ANNs, ANFIS and k-means clustering techniques. J. Hydrol. 2014, 511, 279–289. [Google Scholar] [CrossRef]
  28. Li, J.; Abdulmohsin, H.A.; Hasan, S.S.; Kaiming, L.; Al-Khateeb, B.; Ghareb, M.I.; Mohammed, M.N. Hybrid soft computing approach for determining water quality indicator: Euphrates River. Neural Comput. Appl. 2019, 31, 827–837. [Google Scholar] [CrossRef]
  29. Sengorur, B.; Dogan, E.; Koklu, R.; Samandar, A. Dissolved oxygen estimation using artificial neural network for water quality control. Fresenius Environ. Bull. 2006, 15, 1064–1067. [Google Scholar]
  30. Wen, X.; Fang, J.; Diao, M.; Zhang, C. Artificial neural network modeling of dissolved oxygen in the Heihe River, Northwestern China. Environ. Monit. Assess. 2013, 185, 4361–4371. [Google Scholar] [CrossRef]
  31. Raheli, B.; Aalami, M.T.; El-Shafie, A.; Ghorbani, M.A.; Deo, R.C. Uncertainty assessment of the multilayer perceptron (MLP) neural network model with implementation of the novel hybrid MLP-FFA method for prediction of biochemical oxygen demand and dissolved oxygen: A case study of Langat River. Environ. Earth Sci. 2017, 76, 503. [Google Scholar] [CrossRef]
  32. Chen, D.; Lu, J.; Shen, Y. Artificial neural network modelling of concentrations of nitrogen, phosphorus and dissolved oxygen in a non-point source polluted river in Zhejiang Province, southeast China. Hydrol. Process. 2010, 24, 290–299. [Google Scholar] [CrossRef]
  33. Soyupak, S.; Karaer, F.; Gürbüz, H.; Kivrak, E.; Sentürk, E.; Yazici, A. A neural network-based approach for calculating dissolved oxygen profiles in reservoirs. Neural Comput. Appl. 2003, 12, 166–172. [Google Scholar] [CrossRef]
  34. Schmid, B.H.; Koskiaho, J. Artificial neural network modeling of dissolved oxygen in a wetland pond: The case of Hovi, Finland. J. Hydrol. Eng. 2006, 11, 188–192. [Google Scholar] [CrossRef]
  35. Ranković, V.; Radulović, J.; Radojević, I.; Ostojić, A.; Čomić, L. Prediction of dissolved oxygen in reservoirs using adaptive network-based fuzzy inference system. J. Hydroinform. 2012, 14, 167–179. [Google Scholar] [CrossRef]
  36. Akkoyunlu, A.; Altun, H.; Cigizoglu, H.K. Depth-integrated estimation of dissolved oxygen in a lake. J. Environ. Eng. 2011, 137, 961–967. [Google Scholar] [CrossRef]
  37. Chen, W.B.; Liu, W.C. Artificial neural network modeling of dissolved oxygen in reservoir. Environ. Monit. Assess. 2014, 186, 1203–1217. [Google Scholar] [CrossRef] [PubMed]
  38. Faruk, D.Ö. A hybrid neural network and ARIMA model for water quality time series prediction. Eng. Appl. Artif. Intell. 2010, 23, 586–594. [Google Scholar] [CrossRef]
  39. Han, H.G.; Qiao, J.F.; Chen, Q.L. Model predictive control of dissolved oxygen concentration based on a self-organizing RBF neural network. Control Eng. Pract. 2012, 20, 465–476. [Google Scholar] [CrossRef]
  40. Martí, P.; Shiri, J.; Duran-Ros, M.; Arbat, G.; De Cartagena, F.R.; Puig-Bargués, J. Artificial neural networks vs. gene expression programming for estimating outlet dissolved oxygen in micro-irrigation sand filters fed with effluents. Comput. Electron. Agric. 2013, 99, 176–185. [Google Scholar] [CrossRef]
  41. Antanasijević, D.; Pocajt, V.; Perić-Grujić, A.; Ristić, M. Modelling of dissolved oxygen in the Danube River using artificial neural networks and Monte Carlo Simulation uncertainty analysis. J. Hydrol. 2014, 519, 1895–1907. [Google Scholar] [CrossRef]
  42. Heddam, S. Modeling hourly dissolved oxygen concentration (DO) using two different adaptive neuro-fuzzy inference systems (ANFIS): A comparative study. Environ. Monit. Assess. 2014, 186, 597–619. [Google Scholar] [CrossRef] [PubMed]
  43. Najah, A.; El-Shafie, A.; Karim, O.A.; El-Shafie, A.H. Performance of ANFIS versus MLP-NN dissolved oxygen prediction models in water quality monitoring. Environ. Sci. Pollut. Res. 2014, 21, 1658–1670. [Google Scholar] [CrossRef] [PubMed]
  44. Nemati, S.; Fazelifard, M.H.; Terzi, Ö.; Ghorbani, M.A. Estimation of dissolved oxygen using data-driven techniques in the Tai Po River, Hong Kong. Environ. Earth Sci. 2015, 74, 4065–4073. [Google Scholar] [CrossRef]
  45. Keshtegar, B.; Heddam, S. Modeling daily dissolved oxygen concentration using modified response surface method and artificial neural network: A comparative study. Neural Comput. Appl. 2018, 30, 2995–3006. [Google Scholar] [CrossRef]
  46. Tomić, A.S.; Antanasijević, D.; Ristić, M.; Perić-Grujić, A.; Pocajt, V. A linear and non-linear polynomial neural network modeling of dissolved oxygen content in surface water: Inter- and extrapolation performance with inputs’ significance analysis. Sci. Total Environ. 2018, 610, 1038–1046. [Google Scholar] [CrossRef] [PubMed]
  47. Liu, S.; Xu, L.; Li, D.; Li, Q.; Jiang, Y.; Tai, H.; Zeng, L. Prediction of dissolved oxygen content in river crab culture based on least squares support vector regression optimized by improved particle swarm optimization. Comput. Electron. Agric. 2013, 95, 82–91. [Google Scholar] [CrossRef]
  48. Liu, S.; Xu, L.; Jiang, Y.; Li, D.; Chen, Y.; Li, Z. A hybrid WA–CPSO-LSSVR model for dissolved oxygen content prediction in crab culture. Eng. Appl. Artif. Intell. 2014, 29, 114–124. [Google Scholar] [CrossRef]
  49. Alizadeh, M.J.; Kavianpour, M.R. Development of wavelet-ANN models to predict water quality parameters in Hilo Bay, Pacific Ocean. Mar. Pollut. Bull. 2015, 98, 171–178. [Google Scholar] [CrossRef]
  50. Ravansalar, M.; Rajaee, T.; Ergil, M. Prediction of dissolved oxygen in River Calder by noise elimination time series using wavelet transform. J. Exp. Theor. Artif. Intell. 2016, 28, 689–706. [Google Scholar] [CrossRef]
  51. Fijani, E.; Barzegar, R.; Deo, R.; Tziritis, E.; Konstantinos, S. Design and implementation of a hybrid model based on two-layer decomposition method coupled with extreme learning machines to support real-time environmental monitoring of water quality parameters. Sci. Total Environ. 2019, 648, 839–853. [Google Scholar] [CrossRef]
  52. Huang, Y.; Schmitt, F.G. Time dependent intrinsic correlation analysis of temperature and dissolved oxygen time series using empirical mode decomposition. J. Mar. Syst. 2014, 130, 90–100. [Google Scholar] [CrossRef]
  53. McClelland, J.L.; Rumelhart, D.E. Explorations in Parallel Distributed Processing: A Handbook of Models, Programs, and Exercises; MIT Press: Cambridge, MA, USA, 1989. [Google Scholar]
  54. Haykin, S. Neural Networks and Learning Machines, 3rd ed.; Prentice Hall: Upper Saddle River, NJ, USA, 2009. [Google Scholar]
  55. Kim, S.; Shiri, J.; Kisi, O. Pan evaporation modeling using neural computing approach for different climatic zones. Water Resour. Manag. 2012, 26, 3231–3249. [Google Scholar] [CrossRef]
  56. Kim, S.; Shiri, J.; Kisi, O.; Singh, V.P. Estimating daily pan evaporation using different data-driven methods and lag-time patterns. Water Resour. Manag. 2013, 27, 2267–2286. [Google Scholar] [CrossRef]
  57. Kim, S.; Singh, V.P. Spatial disaggregation of areal rainfall using two different artificial neural networks. Water 2015, 7, 2707–2727. [Google Scholar] [CrossRef]
  58. Deo, R.C.; Sahin, M. An extreme learning machine model for the simulation of monthly mean streamflow water level in eastern Queensland. Environ. Monit. Assess. 2016, 188, 90. [Google Scholar] [CrossRef] [PubMed]
  59. Deo, R.C.; Sahin, M. Forecasting long-term global solar radiation with an ANN algorithm coupled with satellite-derived (MODIS) land surface temperature (LST) for regional locations in Queensland. Renew. Sustain. Energy Rev. 2017, 72, 828–848. [Google Scholar] [CrossRef]
  60. Fahimi, K.; Seyedhosseini, S.M.; Makui, A. Simultaneous competitive supply chain network design with continuous attractiveness variables. Comput. Ind. Eng. 2017, 107, 235–250. [Google Scholar] [CrossRef]
  61. Zounemat-Kermani, M. Assessment of several nonlinear methods in forecasting suspended sediment concentration in streams. Hydrol. Res. 2017, 48, 1240–1252. [Google Scholar] [CrossRef]
  62. Nourani, V.; Baghanam, A.H.; Adamowski, J.; Gebremichael, M. Using self-organizing maps and wavelet transforms for space–time pre-processing of satellite precipitation and runoff data in neural network based rainfall–runoff modeling. J. Hydrol. 2013, 476, 228–243. [Google Scholar] [CrossRef]
  63. Barzegar, R.; Moghaddam, A.A.; Baghban, H. A supervised committee machine artificial intelligent for improving DRASTIC method to assess groundwater contamination risk: A case study from Tabriz plain aquifer, Iran. Stoch. Environ. Res. Risk Assess. 2016, 30, 883–899. [Google Scholar] [CrossRef]
  64. Fahlman, S.E.; Lebiere, C. The cascade-correlation learning architecture. In Advances in Neural Information Processing Systems 2; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1990; pp. 524–532. [Google Scholar]
  65. Kim, S.; Singh, V.P.; Seo, Y. Evaluation of pan evaporation modeling with two different neural networks and weather station data. Theor. Appl. Climatol. 2014, 117, 1–13. [Google Scholar] [CrossRef]
  66. Karunanithi, N.; Grenney, W.J.; Whitley, D.; Bovee, K. Neural networks for river flow prediction. J. Comput. Civ. Eng. 1994, 8, 201–220. [Google Scholar] [CrossRef]
  67. Thirumalaiah, K.; Deo, M.C. River stage forecasting using artificial neural networks. J. Hydrol. Eng. 1998, 3, 26–32. [Google Scholar] [CrossRef]
  68. Mallat, S.G. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 674–693. [Google Scholar] [CrossRef]
  69. Kisi, O.; Cimen, M. A wavelet-support vector machine conjunction model for monthly streamflow forecasting. J. Hydrol. 2011, 399, 132–140. [Google Scholar] [CrossRef]
  70. Nourani, V.; Alami, M.T.; Aminfar, M.H. A combined neural-wavelet model for prediction of Ligvanchai watershed precipitation. Eng. Appl. Artif. Intell. 2009, 22, 466–472. [Google Scholar] [CrossRef]
  71. Kim, S.; Singh, V.P.; Lee, C.J.; Seo, Y. Modeling the physical dynamics of daily dew point temperature using soft computing techniques. KSCE J. Civ. Eng. 2015, 19, 1930–1940. [Google Scholar] [CrossRef]
  72. Seo, Y.; Kim, S.; Kisi, O.; Singh, V.J. Daily water level forecasting using wavelet decomposition and artificial intelligence techniques. J. Hydrol. 2015, 520, 224–243. [Google Scholar] [CrossRef]
  73. Seo, Y.; Kim, S.; Singh, V. Machine learning models coupled with variational mode decomposition: A new approach for modeling daily rainfall-runoff. Atmosphere 2018, 9, 251. [Google Scholar] [CrossRef]
  74. González-Audícana, M.; Otazu, X.; Fors, O.; Seco, A. Comparison between Mallat’s and the ‘à trous’ discrete wavelet transform based algorithms for the fusion of multispectral and panchromatic images. Int. J. Remote Sens. 2005, 26, 595–614. [Google Scholar] [CrossRef]
  75. Kim, S.; Seo, Y.; Rezaie-Balf, M.; Kisi, O.; Ghorbani, M.A.; Singh, V.P. Evaluation of daily solar radiation flux using soft computing approaches based on different meteorological information: Peninsula vs. continent. Theor. Appl. Climatol. 2018, 1–20. [Google Scholar] [CrossRef]
  76. Nason, G. Wavelet Methods in Statistics with R; Springer: New York, NY, USA, 2010. [Google Scholar]
  77. Percival, D.B.; Walden, A.T. Wavelet Methods for Time Series Analysis; Cambridge University Press: New York, NY, USA, 2006; Volume 4. [Google Scholar]
  78. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2014, 62, 531–544. [Google Scholar] [CrossRef]
  79. Luenberger, D.G.; Ye, Y. Linear and Nonlinear Programming, 3rd ed.; Springer: New York, NY, USA, 2008. [Google Scholar]
  80. Deo, R.C.; Şahin, M.; Adamowski, J.F.; Mi, J. Universally deployable extreme learning machines integrated with remotely sensed MODIS satellite predictors over Australia to forecast global solar radiation: A new approach. Renew. Sustain. Energy Rev. 2019, 104, 235–261. [Google Scholar] [CrossRef]
  81. Nash, J.E.; Sutcliffe, J.V. River flow forecasting through conceptual models part I—A discussion of principles. J. Hydrol. 1970, 10, 282–290. [Google Scholar] [CrossRef]
  82. Wilcox, B.P.; Rawls, W.J.; Brakensiek, D.L.; Wight, J.R. Predicting runoff from rangeland catchments: A comparison of two models. Water Resour. Res. 1990, 26, 2401–2410. [Google Scholar] [CrossRef]
  83. Legates, D.R.; McCabe, G.J. Evaluating the use of “goodness-of-fit” measures in hydrologic and hydroclimatic model validation. Water Resour. Res. 1999, 35, 233–241. [Google Scholar] [CrossRef]
  84. Willmott, C.J. On the evaluation of model performance in physical geography. In Spatial Statistics and Models; Springer: Dordrecht, The Netherlands, 1984; pp. 443–460. [Google Scholar]
  85. Willmott, C.J.; Matsuura, K. Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Clim. Res. 2005, 30, 79–82. [Google Scholar] [CrossRef]
  86. Willmott, C.J.; Robeson, S.M.; Matsuura, K. A refined index of model performance. Int. J. Climatol. 2012, 32, 2088–2094. [Google Scholar] [CrossRef]
  87. Chai, T.; Draxler, R.R. Root mean square error (RMSE) or mean absolute error (MAE)?—Arguments against avoiding RMSE in the literature. Geosci. Model Dev. 2014, 7, 1247–1250. [Google Scholar] [CrossRef]
  88. Tsoukalas, L.H.; Uhrig, R.E. Fuzzy and Neural Approaches in Engineering; John Wiley & Sons: New York, NY, USA, 1997. [Google Scholar]
  89. Adamowski, J.; Sun, K. Development of a coupled wavelet transform and neural network method for flow forecasting of non-perennial rivers in semi-arid watersheds. J. Hydrol. 2010, 390, 85–91. [Google Scholar] [CrossRef]
  90. Tiwari, M.K.; Chatterjee, C. Development of an accurate and reliable hourly flood forecasting model using wavelet-bootstrap-ANN (WBANN) hybrid approach. J. Hydrol. 2010, 394, 458–470. [Google Scholar] [CrossRef]
  91. Taylor, K.E. Summarizing multiple aspects of model performance in a single diagram. J. Geophys. Res. Atmos. 2001, 106, 7183–7192. [Google Scholar] [CrossRef]
  92. IPCC. Climate change 2007: The physical science basis. Agenda 2007, 6, 333. [Google Scholar]
  93. Gleckler, P.J.; Taylor, K.E.; Doutriaux, C. Performance metrics for climate models. J. Geophys. Res. Atmos. 2008, 113. [Google Scholar] [CrossRef]
  94. Heo, K.Y.; Ha, K.J.; Yun, K.S.; Lee, S.S.; Kim, H.J.; Wang, B. Methods for uncertainty assessment of climate models and model predictions over East Asia. Int. J. Climatol. 2014, 34, 377–390. [Google Scholar] [CrossRef]
  95. Sigaroodi, S.K.; Chen, Q.; Ebrahimi, S.; Nazari, A.; Choobin, B. Long-term precipitation forecast for drought relief using atmospheric circulation factors: A study on the Maharloo Basin in Iran. Hydrol. Earth Syst. Sci. 2014, 18, 1995–2006. [Google Scholar] [CrossRef]
Figure 1. Structure of standalone soft computing models.
Figure 1. Structure of standalone soft computing models.
Applsci 09 02534 g001aApplsci 09 02534 g001b
Figure 2. Mallat’s discrete wavelet transform (DWT) algorithm for three-level decomposition [73].
Figure 2. Mallat’s discrete wavelet transform (DWT) algorithm for three-level decomposition [73].
Applsci 09 02534 g002
Figure 3. Flowchart for the variational mode decomposition (VMD) algorithm [73].
Figure 3. Flowchart for the variational mode decomposition (VMD) algorithm [73].
Applsci 09 02534 g003
Figure 4. Flowchart for DWT- and VMD-based soft computing modeling.
Figure 4. Flowchart for DWT- and VMD-based soft computing modeling.
Applsci 09 02534 g004
Figure 5. Location of the study area and the water quality monitoring station.
Figure 5. Location of the study area and the water quality monitoring station.
Applsci 09 02534 g005
Figure 6. Comparison of observed and predicted DO values using standalone models for the testing phase.
Figure 6. Comparison of observed and predicted DO values using standalone models for the testing phase.
Applsci 09 02534 g006aApplsci 09 02534 g006b
Figure 7. Scatter plots of observed and predicted DO values using standalone models for the testing phase.
Figure 7. Scatter plots of observed and predicted DO values using standalone models for the testing phase.
Applsci 09 02534 g007
Figure 8. Error residual values for the standalone models for the testing phase.
Figure 8. Error residual values for the standalone models for the testing phase.
Applsci 09 02534 g008
Figure 9. Original water temperature and sub series (details (D1 and D2) and approximation (A2) components) decomposed using Symmlet-6 (S6) mother wavelet.
Figure 9. Original water temperature and sub series (details (D1 and D2) and approximation (A2) components) decomposed using Symmlet-6 (S6) mother wavelet.
Applsci 09 02534 g009aApplsci 09 02534 g009b
Figure 10. Scatter plots of observed and predicted DO values using DWT-based soft computing models for the testing phase.
Figure 10. Scatter plots of observed and predicted DO values using DWT-based soft computing models for the testing phase.
Applsci 09 02534 g010aApplsci 09 02534 g010b
Figure 11. Original water temperature and intrinsic mode functions (IMF1, IMF2, IMF3, and IMF4) decomposed using the VMD algorithm (K = 4 and α = 10).
Figure 11. Original water temperature and intrinsic mode functions (IMF1, IMF2, IMF3, and IMF4) decomposed using the VMD algorithm (K = 4 and α = 10).
Applsci 09 02534 g011aApplsci 09 02534 g011b
Figure 12. Scatter plots of observed and predicted DO values using VMD-based soft computing models for the testing phase.
Figure 12. Scatter plots of observed and predicted DO values using VMD-based soft computing models for the testing phase.
Applsci 09 02534 g012
Figure 13. Taylor diagrams for evaluating the performance of the applied models.
Figure 13. Taylor diagrams for evaluating the performance of the applied models.
Applsci 09 02534 g013
Figure 14. Violin plots of observed and predicted DO concentration.
Figure 14. Violin plots of observed and predicted DO concentration.
Applsci 09 02534 g014
Table 1. Summary of statistics parameters of input and output variables (n = 232).
Table 1. Summary of statistics parameters of input and output variables (n = 232).
VariableUnitMinMaxMedianMeanSDCVCC
InputClmg/L40.000850.000200.000247.649164.9730.6660.418
NOxmg/L0.0010.4600.0720.0890.0870.9780.554
TDS 134.0001950.000559.500642.362355.1630.5530.394
pHStandard Units6.1008.2207.1807.1480.4210.0590.760
WT°C8.98032.22023.510232.3785.2110.022−0.544
OutputDOmg/L0.09011.4805.9205.4852.5650.4681.000
Note: CV: the coefficient of variation, and CC: the coefficient correlation between inputs and DO.
Table 2. Input combinations for structuring the MLP and CCNN models.
Table 2. Input combinations for structuring the MLP and CCNN models.
ID of the Input CombinationInputInput No.Output
ICl, NOx, TDS, pH, WT5DO
IICl, NOx, pH, WT4DO
IIINOx, pH, WT3DO
IVWT, NOx2DO
VWT1DO
VIPH1DO
Table 3. Performance statistics of various input combination using MLP and CCNN models.
Table 3. Performance statistics of various input combination using MLP and CCNN models.
ModelIDTopologyTraining PhaseTesting Phase
RMSE (mg/L)NSEWIR2MAE (mg/L)RMSE (mg/L)NSEWIR2MAE (mg/L)
MLPI5-2-11.211 *0.778 *0.933 *0.781 *0.924 *1.4250.6630.8970.7691.126
II4-2-11.2370.7680.9300.7690.9381.3590.6940.9020.7741.062
III **3-4-11.3060.7420.9180.7520.9941.261 *0.736*0.919 *0.801 *0.989 *
IV2-4-11.8300.4930.8170.4961.4371.4270.6620.8680.7071.072
V1-1-12.1380.3080.6760.3101.8102.0070.3320.6840.3351.633
VI1-1-11.5850.6200.8710.6201.2642.1380.2420.7410.5681.773
CCNNI5-1-11.2270.7720.9330.7730.9531.1720.7720.9350.7950.790
II4-1-11.196 *0.783 *0.937 *0.785 *0.881 *0.5810.7730.9350.8010.195
III **3-2-11.2450.7650.9310.7670.9220.550 *0.797 *0.942 *0.825 *0.185 *
IV2-3-11.7600.5310.8370.5311.3410.6740.6950.8960.7050.246
V1-10-11.8160.5010.8060.5021.4600.9260.4240.7980.4520.385
VI1-2-11.5790.6220.8750.6241.2510.7570.6150.8690.6540.297
Note: * shows the best performance for each column; ** stands for introducing the best model.
Table 4. Performance statistics using DWT-MLP-III and DWT-CCNN-III models.
Table 4. Performance statistics using DWT-MLP-III and DWT-CCNN-III models.
ModelDWTTopologyTraining PhaseTesting Phase
RMSE (mg/L)NSEWIR2MAE (mg/L)RMSE (mg/L)NSEWIR2MAE (mg/L)
MLP-IIIC69-27-10.5230.9580.9890.9580.3450.3640.9780.9940.9780.291
C129-26-10.4180.9740.9930.9740.3140.1780.9790.9950.9790.066
C189-21-10.4410.9710.9920.9710.3430.2020.9730.9930.9750.072
D69-22-10.4370.9710.9930.9710.3100.1610.9830.9960.9830.061
D129-22-10.3460.9820.9950.9820.2550.1800.9780.9950.9790.066
D189-28-10.4470.9700.9920.9700.3360.2230.9670.9920.9700.086
S69-22-10.4370.9710.9930.9710.3100.1610.9830.9960.9830.061
S129-28-10.4200.9730.9930.9730.3080.1840.9770.9940.9770.069
S189-37-10.4340.9710.9930.9720.2940.1860.9770.9940.9770.068
CCNN-IIIC69-0-11.3410.7270.9140.7301.0351.3600.6930.9070.7391.055
C129-0-11.3350.7300.9160.7351.0410.6560.7110.9130.7470.250
C189-1-11.2170.7760.9350.7800.9420.6110.7490.9270.7940.219
D69-2-11.2160.7760.9340.7780.9560.6050.7540.9280.7920.219
D129-1-11.2450.7650.9290.7680.9710.6590.7080.9180.7570.244
D189-0-11.3300.7320.9170.7351.0280.8690.7090.9120.7550.245
S69-2-11.2160.7760.9340.7780.9560.6050.7540.9280.7920.219
S129-0-11.3360.7300.9150.7341.0460.6830.6870.9030.7280.258
S189-1-11.2560.7610.9290.7630.9580.6120.7490.9280.7870.207
Table 5. Performance statistics using VMD-MLP-III and VMD-CCNN-III models.
Table 5. Performance statistics using VMD-MLP-III and VMD-CCNN-III models.
ModelVMDTopologyTraining PhaseTesting Phase
RMSE (mg/L)NSEWIR2MAE (mg/L)RMSE (mg/L)NSEWIR2MAE (mg/L)
MLP-IIIK = 3, α = 59-22-10.3540.9810.9950.9810.2700.3590.9790.9950.9790.280
K = 4, α = 512-25-10.3350.9830.9960.9830.2400.1460.9860.9960.9860.054
K = 4, α = 1012-19-10.2570.9900.9970.9900.1910.1070.9920.9980.9930.034
CCNN-IIIK = 3, α = 59-0-11.3120.7390.9200.7441.0031.3070.7170.9160.7620.965
K = 4, α = 512-0-11.3220.7350.9210.7371.0260.6110.7500.9260.7920.228
K = 4, α = 1012-0-11.3270.7330.9210.7381.0360.5970.7610.9290.7910.218

Share and Cite

MDPI and ACS Style

Zounemat-Kermani, M.; Seo, Y.; Kim, S.; Ghorbani, M.A.; Samadianfard, S.; Naghshara, S.; Kim, N.W.; Singh, V.P. Can Decomposition Approaches Always Enhance Soft Computing Models? Predicting the Dissolved Oxygen Concentration in the St. Johns River, Florida. Appl. Sci. 2019, 9, 2534. https://doi.org/10.3390/app9122534

AMA Style

Zounemat-Kermani M, Seo Y, Kim S, Ghorbani MA, Samadianfard S, Naghshara S, Kim NW, Singh VP. Can Decomposition Approaches Always Enhance Soft Computing Models? Predicting the Dissolved Oxygen Concentration in the St. Johns River, Florida. Applied Sciences. 2019; 9(12):2534. https://doi.org/10.3390/app9122534

Chicago/Turabian Style

Zounemat-Kermani, Mohammad, Youngmin Seo, Sungwon Kim, Mohammad Ali Ghorbani, Saeed Samadianfard, Shabnam Naghshara, Nam Won Kim, and Vijay P. Singh. 2019. "Can Decomposition Approaches Always Enhance Soft Computing Models? Predicting the Dissolved Oxygen Concentration in the St. Johns River, Florida" Applied Sciences 9, no. 12: 2534. https://doi.org/10.3390/app9122534

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop