Streamflow forecasting is highly crucial in the domain of water resources. For this study, we coupled the Wavelet Transform (WT) and Artificial Neural Network (ANN) to forecast Gilgit streamflow at short-term (T0.33 and T0.66), intermediate-term (T1), and long-term (T2, T4, and T8) monthly intervals. Streamflow forecasts are uncertain due to stochastic disturbances caused by variations in snow-melting routines and local orography. To remedy this situation, decomposition by WT was undertaken to enhance the associative relation between the input and target sets for ANN to process. For ANN modeling, cross-correlation was used to guide input selection. Corresponding to six intervals, nine configurations were developed. Short-term intervals performed best, especially for T0.33; intermediate intervals showed decreasing performance. However, interestingly, performance regains back to a decent level for long-term forecasting. Almost all the models underestimate high flows and slightly overestimate low- to intermediate-flow conditions. At last, inference implicitly implies that shorter forecasting benefits from extrapolated trends, while the good results of long-term forecasting is associated to a larger recurrent pattern of the Gilgit River. In this way, weak performance for intermediate forecasting could be attributed to the insufficient ability of the model to capture either one of these patterns.

  • Good, decent, and low performances were observed corresponding to short, long, and intermediate forecasting.

  • Models underestimate at high flow and overestimate at low to intermediate flow conditions, implying a limited sample for model training.

  • Short-term forecasting follows short-term trends, while long-term follows larger recurrent patterns probably associated with snow accumulation and melting.

Accurate streamflow forecasting is of instrumental importance in the domain of water resources, especially in the sub-disciplines of flood control, water supply, hydropower, reservoir operation, etc. (Senthil Kumar et al. 2013; Nacar et al. 2017). The streamflow generation is dictated by a number of parameters that simultaneously interact at varying spatio-temporal scales. These include the climate variables (precipitation, temperature, evapotranspiration, etc.), the ground conditions (land use, roughness, etc.), radiation, anthropogenic activities, etc. (Nacar et al. 2017). All these factors make flow processes highly non-linear and complex. The researchers, faced with this bewildering complexity of nature, have resorted to the development of streamflow forecasting models that can adequately simulate the flow dynamics suitably and with sufficient accuracy (Kalteh 2013).

Typically, modeling methodologies are broadly divided into two categories: process-based models and data-driven models (Patel & Joshi 2017; Zhang & Yang 2018). Process-based modeling assumes the river basin processes as interlinked mathematical functions (Islam 2011; Hassan et al. 2014). The process-based models are subdivided into conceptual and physically based models. Conceptual models are mostly based on simplified assumptions and hence their utility is limited to simple uses, while physically based models can incorporate complex processes (Islam 2011). These types of models, generally, require a higher number of data inputs, such as climate, physiographic inputs, etc., and their processing is based on the holistic consideration of these inputs for modeling. Physiographic inputs such as land use and vegetation are generally derived from the satellite via indirect measurements and could contain uncertainty (Hosseini et al. 2022). Land use/cover is also incorporated as a single layer and therefore, considered time-stationary, which means that it is assumed that during modeling the land use dynamics (monthly and annual variations) will not affect the flow generation. These limitations along with the paucity of data are adjusted by extensive alterations in other model parameters during calibration, which results in an improper model parameterization and hence degrades the quality of the simulation (Nesterova et al. 2021). Secondly, physical models necessarily require future weather data for forecasting, which is an issue in itself. Until now, generally, weather forecasting accuracy has been 80% for a 7-day forecast and 50% for a 10-day forecast (Scijinks n.d). Nevertheless, predicting quantity is even more challenging (precipitation, etc.) (Šaur 2015). Therefore, in view of the aforementioned limitations (in the model and weather forecast), forecasting using physical modeling is an onerous task that is prone to high uncertainty. Alternatively, ensemble-based modeling could be a potent option as compared to a standalone model, as it could define probabilistic ranges against a forecasting sequence. However, the development and calibration of ensembles are computationally intensive and rigorous tasks.

Besides physically based models, Artificial Neural Network (ANN) black box models are another popular choice for hydrological modeling. ANNs are parallel-distributed processing systems analogous to the biological neural networks of the human brain (ASCE 2000; Govindaraju 2000). Due to this, these models are robust in approximating intricate issues such as approximating non-linear relations (Patel & Joshi 2017), pattern recognition (Jianjin et al. 2017), classification, prediction, and sequential analysis (Akhtar et al. 2009). Several types of ANN models are in use for water resources applications, such as Hayder et al. (2022) implementation of Non-Linear Autoregressive with Exogenous Inputs (NARX) and Long Short-Term Memory (LSTM) for forecasting river flows. Wunsch et al. (2020) use LSTM, NARX, and a Convolutional Neural Network (CNN) for forecasting groundwater levels. Further, Apaydin et al. (2020) investigate three types of ANNs: NARX, LSTM, and the popular Multi-Layer Perceptron (MLP) for reservoir outflow forecasting. The specialty of ANN models is that they can directly relate inputs to the required target without considering the underlying hydrological processes (Noori & Kalin 2016; Pitta et al. 2016). The ANN model benefits from the affiliated correlation present between the input and target variable and, therefore, for optimum convergence, unlike the physical model, it does not enforce the fulfillment of specific input requirements. Any exogenous or non-exogenous variables with necessary affiliation/correlation could be used for result derivation. This unique attribute of ANN can compensate for physically based forecasting's shortcomings. However, the inclusion of these variables is subject to data availability and access. That's why parsimonious approaches are often more preferred and commonly used worldwide (Kalteh 2013).

The wavelet transforms are physically based signal processing techniques that are recently introduced and are being used in conjunction with ANN and other machine learning approaches for hydrological applications. They deliver the information of the primary signal or data in time and frequency domains, thus providing valuable information regarding their structure and also giving a time-frequency depiction of the primary signal at multiple temporal resolutions (Daubechies 1990). Several studies have used ANN-Wavelet Transform (WT) hybrids for obtaining daily forecasts. For instance, Khan et al. (2021) and Partal & Cigizoglu (2008) used a WT–ANN hybrid for forecasting the daily suspended sediment load. Adamowski & Sun (2010) used ANN and WT for forecasting the flow of non-perennial rivers and reported that hybrid modeling provides better results than the exclusive ANN model. Adamowski & Chan (2011) used the ANN-wavelet model for forecasting groundwater levels. Kalteh (2013) used wavelet and ANN along with other deep learning approaches for forecasting streamflow at Kharjegil and Ponel stations in Northern Iran. Freire et al. (2019) reported that coupling WT–ANN provides better results than standalone ANN for short-term streamflow forecasting. Recently, Wang et al. (2022) coupled several machine learning algorithms with wavelet theory for determining monthly scale river discharge. Qasem et al. (2019) modeled monthly pan evaporation by hybridizing deep learning Support Vector Regression (SVR) and ANN schemes with wavelet transformations. Most of these studies concluded that WT–ANN hybrids are more robust in forecasting than exclusively ANN, which informs that WT–ANN hybridization possesses a strong harnessing ability that can leverage model performance.

The Gilgit Basin, being part of the Upper Indus Basin (UIB), is characterized by high-altitude and cold terrain. As one of the largest tributaries of the UIB, it provides crucial input into the Indus water system that tends to regulate the downstream water fluctuations (flooding and droughts). Physiographically, a large portion of the basin remains covered with snow all year long (Adnan et al. 2016a), which signifies that the snow-melting process regulates the annual perennial flow cycles of the Gilgit River. However, the melting process is convolutionally uneven and heavily relies upon numerous intricate parameters (precipitation distribution and amount, temperature distribution, rain on snow, snow density, atmospheric pressures, solar radiation, surface canopy, hill shades, etc.). Previously, most studies on the Gilgit Region were based on simplified physical assumptions (e.g., Degree Day) that only used a few generic parameters (precipitation, temperature, etc.) for imitating the melting process, which made either the results or the model setting ambiguous (Bashir & Rasul 2010; Adnan et al. 2016b; Latif et al. 2019). Nevertheless, these studies were relevant to streamflow prediction/hydro-climatic modeling rather than forecasting. No works of literature were found that directly addressed streamflow forecasting in our region of interest. Unsatisfactory meteorological information in the region is another constraint that creates a complication for streamflow forecasting. Various studies have reported a paucity of hydrometeorological stations and other inherent deficiencies at UIB (Syed et al. 2022). Further, stochastic variation in precipitation gradient at a limited spatial scale also adds up to uncertainty in flow simulation that obscures the streamflow generation process (Schreiner-Mcgraw & Ajami 2020). Faced with limitations, we presume that ANN would be a better option for flow forecasts, as it directly relates an input to an output, skipping the underlying intricate physical processes and solving based on a given relation between these variables. Further, lags due to snowmelt and other unknown reasons can also be adjusted by simply applying time-lag adjustments to inputs. Further, the wavelet transforms can be used in conjunction with ANN, as it allows to decomposition of the signals to multiple degrees, which could remove unwanted noise and clarify the most associative hidden signals. In this way, unlike standalone ANN, coupled WT–ANN could be better at tracking the exposed signals and calculating the flow magnitudes in an appropriate way.

The objective of this study is to forecast the Gilgit River's flow using WT and ANN hybrids at several forecasting intervals (short, intermediate, and long-terms). It will comprehensively discuss each phase of the model's implementation and highlight internal and external strengths and limitations.

The Gilgit River Basin (Figure 1) lies between latitudes 35°46′05 N and 36°51′16″N and longitudes 72°25′02″E and 74°19′25 E. It is located in the extreme north of Pakistan in the foothills of the Himalaya, Karakorum, and Hindukush (HKH) mountain ranges. The Gilgit River originates from the Shandor Lake and joins the Indus Basin at the Bunji/Partab Bridge at Jaglot Town, Baltistan, Pakistan. The basin area is approximately 12,648 km2 above the river gauge station. The topography of the basin is characterized by high-altitude, cold, and rugged terrain. It has a mean elevation of 3,992 m above sea level having 923 known glaciers in the area, covering an area of 858.168 km2 (Cogley 2010; Adnan et al. 2016b). The valleys, situated at lower elevations, are characterized as arid, while those at higher elevations are semi-arid. A maximum of 1,000 mm of rainfall is received by the southern part of the catchment, while the maximum temperature in high-altitude valleys in the Gilgit catchment could range from 10 to 15 °C (Adnan et al. 2016a). The average annual discharge at the gauge point from 1998 to 2015 was 295.20 m3/s. Major tributaries of the Gilgit River are the Handrab, Langar, Yasin, Phandar, and Ishkoman Rivers.
Figure 1

Location map of the Gilgit River Basin.

Figure 1

Location map of the Gilgit River Basin.

Close modal

Wavelet transform

Transformations are mathematical functions that are used to gather signal information, represented by the frequency with a certain magnitude. Among several transformation functions, Fourier transforms are the most popular. The Fourier transform can process stationary signals (constant frequency for a long time) but cannot provide frequency information of non-stationary or complex signals. The Window Fourier Transform (WFT) addresses this frequency localization (what frequency at what time instants) problem by processing the frequency using a localized, time-based window. The window slides over time, assuming the portion of the non-stationary signal inside the window as a stationary signal and calculates the Fast Fourier Transform (FFT) at each instant. However, despite the improvement, its accuracy highly depends upon its window function, which affects its frequency resolution. For instance, if we consider fixing the window size when sliding over a non-stationary signal, frequency localization could be lost if the frequency inside the window is either too low or too high (Santos & Silva 2014).

The WT addresses the above WFT concern by examining the signal in different frequencies and resolutions (Santos & Morais 2013; Santos & Silva 2014; Honorato et al. 2019; Santos et al. 2019). Analogous to the window function, the WT uses scalable short waves or mother waves as a sliding function, however, in this case, a wavelet is used to overlap the raw signal. The wavelet can be expanded and dilated by changing the scalar function as it moves across the signal. This is done so that the high frequencies of the wavelet can capture the high frequencies of the signal and the low frequencies of the wavelet can capture the low frequencies of the signal (Santos et al. 2019). The general equation of WT is given in the following (Santos et al. 2019):
(1)
where is the scaler parameter, is a translation parameter, is signal, is time, is a short wavelet or mother wavelet and denotes complex conjugate.
The Equation (1) also represents Continuous WT (CWT). Process signal using CWT is computationally intensive and requires a large amount of data (Kalteh 2013). Alternatively, Discrete WT (DWT), which is obtained through discretizing of and translation parameters of the CWT is less computationally intensive and easy to implement. The DWT equation is given in the following equation (Santos et al. 2019):
(2)
where and are integers that control the scale and time, is the time, is a specified fixed dilation step > 1, and is the location parameter that must be > 0. The term in the above equation normalizes the functions.
In DWT, the main signal (Q) passes through two filters, named, high-pass and low-pass filters. A low-pass filter lowers the amount of information carried by the signal, which is called Approximations (A) and a high-pass filter represents the lost information called, Detailed (D) (Santos & Silva 2014). In physical terms, the approximation coefficients represent the low-frequency component of the signal, while the detail coefficients represent the high-frequency component. To put it another way, approximation coefficients capture the overall trend or general behavior of a signal, whereas detail coefficients capture local variations or fluctuations. The original signal can be reconstructed by simply adding up approximations and details. The approximation is the most important component as it holds the identity of the signal (Santos & Silva 2014). The approximation can be further decomposed into sub-approximations and sub-details and so on, as explained in the following equations:
(3)
(4)

Artificial neural network

An ANN is a computational approach inspired by a biological neural system. It uses computing units called artificial neurons in the form of a layered network to map out complex relations between given input and target data. Each neuron in the network is interlinked with the neuron of the neighboring layer. ANN has the ability to process co-evolutionary relations by skipping the intermediate physical processes. Fundamentally, ANN is comprised of three components: the input, hidden, and the output layers. The input layer and output layer consist of an equal number of elements. In the simplest neural network, where the network consists of a single layer hidden neuron, the normalized input features from the input layer are multiplied with weights (w). The weighted sum of these inputs is then added to bias (b) to form a linear function. Weight and bias are typical, random parameters ranging from −1 to 1 and depicting excitatory and inhibitory signals, respectively. Weight regulates the mathematical relationship between target and input variables. For dynamic output, a linear function is converted into a non-linear function by applying activation or transfer functions. An activation function is a non-linear function that is differentiable with respect to weight and bias. The basic equation of a single neuron is expressed as follows:
(5)
where is the resultant, f is the activation function of the neuron, and X is the ith input element (where i = 1,. . n).
The most common ANN structure for hydrological modeling applications is the feed-forward MLP (Seydou et al. 2010; Kalteh 2013; Ali & Shahbaz 2020). It normally consists of one to two hidden layers. Each neuron in the hidden layer acts as an input to a neuron of the next layer and so forth. In the end, the output layer consists of a single neuron that computes the output (Jimeno-Sáez et al. 2018; Hosseini et al. 2022). The activation functions decide the degree of activation of each neuron of the layer. The activation function for each layer is typically customizable, and each one can choose a different type of activation function. Among several activation functions, the Rectified Linear Unit (ReLU) is the most commonly used. It is computationally less intensive than other popular functions such as sigmoid and tanh (Xu et al. 2015; Ali & Shahbaz 2020). Mathematically, it is given as follows:
(6)

The back-propagation algorithm is a supervised learning technique that is used for error reduction in MLP. It works in forward and backward passes. In the forward pass, inputs are passed through the intervening layers to the output layer, where losses are calculated by means of the loss function (model residual of target and output). Errors are then back-propagated or back passed for adjustment. In this way, the process repeatedly adjusts the weights (which were initially randomly assigned) at every iteration (Jimeno-Sáez et al. 2018; Ali & Shahbaz 2020; Hosseini et al. 2022). This process keeps iterating until the following rules are met: epochs reach the set limit, or the gradient becomes 0, or the model fulfills early stopping conditions. Previously, Aqil et al. (2007) attested several training algorithms (e.g., Levenberg–Marquardt (LM) back-propagation, Bayesian regularization, and gradient descent with momentum) and found the LM back-propagation to be more effective than other training algorithms. Further, studies also reported LM's better performance and a faster convergence rate as compared to other popular back-propagation algorithms such as conjugate gradient and resilient back-propagation (Kisi & Uncuoğlu 2005; Tabbussum & Qayoom 2020).

Statistical evaluators

This study considered four commonly used statistical evaluators or metrics, namely, Pearson Correlation (R), Coefficient of Determination (R2), Nash–Sutcliffe Efficiency (NSE), and Root Mean Square Error (RMSE), also shown in Table 1. Pearson Correlation (R) explains the strength and direction of relationships. Its values range from −1 to 1. The value −1 indicates a perfectly inverse relation, while 1 represents a perfectly proportional relation; nearly zero or zero shows poor or no relation at all. The R2 is the square of the Pearson Correlation that highlights the collinearity or strength of the relationship in terms of a positive number. It is a common metric for measuring the goodness of fit of the models. The least value 0, represents no relation, while the maximum value 1, represents a perfect fit. Similar to R2, NSE is nearly identical to R2 except it quantifies model simulation. Its value ranges from -∞ to 1. The NSE value can also be written in terms of RMSE, as given in Table 1. According to the relationship, if RMSE is equal to the standard deviation of observation, then NSE will be 0, while if the RMSE value is half of the standard deviation of observation, then NSE will be 0.75. As such, lower RMSE values are highly desirable.

Table 1

Statistical evaluator and details

Obj. functionsEquationsRange
Coefficient of Determination (R2 (0, 1) 
R  (1, 1) 
NSE  (-, 1) 
RMSE  (0,
Obj. functionsEquationsRange
Coefficient of Determination (R2 (0, 1) 
R  (1, 1) 
NSE  (-, 1) 
RMSE  (0,

Note: N, the number of flow values; Qobs and Qsim, the observed and simulated flows for ith observations; Q*obs and Q*sim, the means of the observed and simulated flow values; Sobs and Ssim, the standard deviation of the corresponding observed and simulated flows.

Model setup

For this study, 10-daily temporal resolution stream flow from 1999 to 2015 was used as the main input and output, comprising of in total of 612 data points. The 10-daily is a moderate temporal resolution commonly used by engineering practitioners, where daily records are either unavailable or ambiguous. It is more suitable than commonly used daily and monthly resolutions since it smooths out the daily noise and at the same time preserves sub-monthly trends. The first 10-daily represents the mean of the first 10 days of the month; the second 10-daily represents the mean from 11th to 20th days of the month and the third 10-daily represents the mean of the rest of the remaining days in a month. Flow was selected as an exclusive predictor. Exogenous climate variables (precipitation, temperature, etc.) were not taken into account in this study as they are dependent upon several factors such as data availability, public accessibility, and reliability (errors in data; wind-induced, random, etc.). As explained earlier in the introduction, the Gilgit River Basin has deficient and biased observation climatology, which could degrade the model simulation; therefore, we opted to omit the inclusion of exogenous climate variables. Further, we have found no explicit guidelines regarding the selection of forecasting intervals in the literature; however, we opted to adopt an interval range as a multiple of two rather than taking constant intervals. In this way, inferences about larger ranges can be computed by only considering a few intervals. Additionally, it is also allowed to estimate inferences on sub-intervals via interpolation with reasonable accuracy. Following, monthly intervals were selected for forecasting based on 10-daily data (T0.33; 0.33-month, T0.66; 0.66-month, T1; 1-month, T2; 2-month, T4; 4-month, and T8; 8-month). Further, we categorize intervals as short-term (T0.33 and T0.66), intermediate-term (T1), and long-term (T2, T4 , and T8) forecasting intervals.

In total, two types of input variables were considered for the ANN modeling, namely, flow and wavelet-based. The flow-based included the stream flow (Q), while the wavelet-based comprised Approximation (A) and Details (D). Monthly interval flow-based input variables (QT) are as follows: QT0.33, QT0.66, QT1, QT2, QT4, and lastly QT8. For the wavelet-based input variables, stream flow (Q) was decomposed using the Daubechies wavelet setting. This resulted in first-level Approximation (A1) and Details (D1). The A1 was then further decomposed to give A2 and D2, and this procedure was then repeated until obtaining A10 and D10 (subscript represents the level of decomposition). At the end, the series were adjusted for the given monthly intervals. The input, monthly interval Approximations (T) and Details (T) are as follows: Approximations, An = [T0.33, T0.66, T1, T2, T4, and T8]; Details, Dn = [T0.33, T0.66, T1, T2, T4, and T8], where n = (1,…, 10) represented the number of level of decomposition. Further, cross-correlation analysis using Pearson Correlation was performed between the input variable and stream flow to find the highly correlated variables. This analysis was used for good effect as highly correlated variables were selected for developing the input configurations for the ANN modeling. The configurations were selected based on two criteria; firstly, the best variables were selected considering the same interval. Secondly, the best variables were selected from the other intervals, provided that the interval was less than or equal to the reference interval. For instance, for 4 months streamflow forecasting, only (AT4, AT8, QT4, QT8, etc.) can be used as input variables and no other variables with T < 4 will be considered eligible. By doing so, the model was able to make use of good relationships existing at large intervals.

The development of an ANN model requires determining the optimum configuration of neurons and layers in the model. The number of neurons in the input and output layers of the model is equal to the number of input and output variables. However, generally, the number of neurons in the intervening layers is adjusted by trial-and-error. Our study opted for the following settings, preferring parsimonious conditions: input layer with four neurons denoting a maximum of 4 inputs from streamflow, for the hidden layer and 1 single layer with 10 neurons was selected after trial-and-error and 1 neuron for the output layer denoting stream flow output. For better performance and rapid processing, the ReLU transfer function was selected. The ANN models were randomly trained using a typical 70% of the data and then tested on 30% of the remaining data. Early stopping condition was considered for avoiding overfitting. LM was used as a back-propagation algorithm for better results and faster convergence. The models were statistically evaluated using NSE, RMSE, and CC (Correlation Coefficient). Details are provided in Section 3.3.

Cross-correlation analysis

A comprehensive cross-correlation analysis was carried out between the flow and the input variables, shown in Table 2. Synoptically, correlation reduces as forecasting intervals increase. However, interestingly, it was also found that for T4 and T8 intervals, correlation slightly improved in terms of negative correlation, which might be indicating an inverse transformation of signals. Nevertheless, this type of behavior also helped model learning and assimilation of data.

Table 2

Cross-correlation analysis of flow and input variables

 
 

Analysis showed that in the case of flow, Q, a strong correlation was achieved corresponding to the short-term intervals (T0.33 and T0.66) followed by the intermediate interval T1, which also fell in a fair range. However, signal correlation drastically reduced to a fairly low value for the T2 interval, though minor improvement was observed for the long-term (T4 and T8) intervals. In the case of approximations, 10 decomposed approximation series were compared to the flow at different intervals, and a correlation was obtained. The results showed that the A1 and A2 series were among the best performers, as they were able to preserve most of the information and yield the maximum correlation with the flow. The Details series depicted a relatively lower correlation with the flow, nevertheless, the pattern remained the same as for the approximation series.

Forecasting performance

There are no explicit guidelines regarding the number of inputs for an ANN model, however, it should be carefully selected as the input variables are instrumental in mapping the underlying non-linear processes. More than necessary inputs can reduce the learning rate of ANNs while a deficient number of inputs could result in the formation of a weak model (Jimeno-Sáez et al. 2018). In our case, four highly correlated variables from the cross-correlation analysis were used to prepare each of the nine input configurations. Additional details and criteria for developing input configuration are provided in Section 3.4 of this article.

A statistical summary of the comparison between observed and input configurations of different monthly forecasting intervals is presented in Table 3. For inspecting flow variations and discrepancies, the flow series are depicted in Figure 2. Statistical metrics were calculated on predicted training and testing sets. Holistically, obtained NSE and R2 values have yielded almost analogous results, indicating lower discrepancies in trained model and simulation values. This can be more closely inspected by analyzing Supplementary material, Figure S1. Additionally, some configuration scores appear to be slightly better in the testing phase, which may indicate that models are more likely to predict most of the testing values or a certain range of values.
Table 3

Configuration details and statistical summary of forecasting intervals

LabelsaForecasting intervalsConfigurationsTraining
Testing
R2NSERMSER2NSERMSE
a1 Short-term: T0.33 Q T0.33, A1 T0.33, A2 T0.33, D1 T0.33 0.87 0.87 125.26 0.89 0.88 107.22 
a2 QT0.33, A1 T0.33, A2 T0.33, A1 T0.66 0.87 0.87 124.93 0.88 0.88 107.65 
b1 Short-term: T0.66 QT0.66, A1 T0.66, A2 T0.66, D2 T0.66 0.79 0.78 159.91 0.78 0.77 150.11 
c1 Intermediate: T1 QT1, A1 T1, D1 T1, D2 T1 0.70 0.68 193.79 0.69 0.68 178.26 
d1 Long-term: T2 QT2, A1 T2, D1 T2, D2 T2 0.72 0.71 183.98 0.78 0.78 147.69 
d2 Q T4, Q T8, A1 T4, A2 T8 0.73 0.73 179.22 0.73 0.66 184.31 
e1 Long-term: T4 QT4, A1 T4, A2 T4, D1 T4 0.74 0.74 175.92 0.79 0.76 153.60 
e2 Q T4, Q T8, A1 T4, A2 T8 0.73 0.73 179.44 0.74 0.66 184.10 
f1 Long-term: T8 QT4, A1 T8, A2 T8, D1 T8 0.76 0.76 167.99 0.81 0.79 142.35 
LabelsaForecasting intervalsConfigurationsTraining
Testing
R2NSERMSER2NSERMSE
a1 Short-term: T0.33 Q T0.33, A1 T0.33, A2 T0.33, D1 T0.33 0.87 0.87 125.26 0.89 0.88 107.22 
a2 QT0.33, A1 T0.33, A2 T0.33, A1 T0.66 0.87 0.87 124.93 0.88 0.88 107.65 
b1 Short-term: T0.66 QT0.66, A1 T0.66, A2 T0.66, D2 T0.66 0.79 0.78 159.91 0.78 0.77 150.11 
c1 Intermediate: T1 QT1, A1 T1, D1 T1, D2 T1 0.70 0.68 193.79 0.69 0.68 178.26 
d1 Long-term: T2 QT2, A1 T2, D1 T2, D2 T2 0.72 0.71 183.98 0.78 0.78 147.69 
d2 Q T4, Q T8, A1 T4, A2 T8 0.73 0.73 179.22 0.73 0.66 184.31 
e1 Long-term: T4 QT4, A1 T4, A2 T4, D1 T4 0.74 0.74 175.92 0.79 0.76 153.60 
e2 Q T4, Q T8, A1 T4, A2 T8 0.73 0.73 179.44 0.74 0.66 184.10 
f1 Long-term: T8 QT4, A1 T8, A2 T8, D1 T8 0.76 0.76 167.99 0.81 0.79 142.35 

Note: The best configurations are highlighted in bold.

aThe objects in labels (column) referred to the results of the simulations in Figure 2.

Figure 2

Observation and simulated forecasting flow series for each interval.

Figure 2

Observation and simulated forecasting flow series for each interval.

Close modal

Individual inspection shows that two configurations of the short-term (T0.33) forecasting interval stand out in terms of performance, both in training and testing epochs. Both configurations adequately fit into observation data, as can be seen in Figure 2(a1-2). The given configuration against short-term (T0.66) also yielded good and consistent performance, as depicted in Figure 2(b1). However, a significant drop in metric values was observed compared to the latter case. For intermediate intervals, the model produced comparatively low but moderate inferences. In Figure 2(c1), discrepancies under high flow conditions could be observed. Interestingly, a relatively fair performance was recorded in the case of the long-term T2 interval, considering the poor correlation of its associated variables. This might be either due to the unique ability of ANN to track hidden links of poorly related variables or else the weak variables, when in combination, were able to show a higher correlation value. The later hypothesis can be further explained by conjecturing that the multiple weak variables in an ensemble formation might have raised the signal strength, which allowed the generation of a better signal and supported ANN to assimilate appropriately. Nevertheless, the model poorly produced high flow conditions Figure 2(d1-2). In the case of long-term (T4) prediction, the model showed improved results as compared to the T2 interval. This was most probably because of improved correlation due to the inverse transformation of stream signals, as indicated by an improved negative correlation in Table 2. This might be an indication of the influence of annual recurrent pattern of the Gilgit River on long-term forecasts. Figure 2(e1-2) displays the T4 flow series and it can be seen that its performance is below par in predicting extreme flow conditions. Likewise, at 8-month time interval, T8 the model performance was suboptimal as shown in Figure 2(f1). Overall, given the fact that the training and testing scores are comparable (in some cases slightly better), it is clear that the model's lack of ability to predict extreme values represents a major hurdle in achieving a perfect score.

Performance at different flow conditions

Forecasting performance was evaluated at different instants of time using a Flow Duration Curve (FDC). FDC was discretized into three sets of exceedance probability based on flow conditions as follows: high flow, , intermediate flow, and low flow, .

The results in Figure 3 suggest that WT–ANN, as expected, performed well at short- and intermediate-term intervals (T0.33, T0.66, and T1). Beyond that, the analysis depicted irregular flow conditions, implying abrupt rises, and exaggerated magnitudes. Almost all the forecasting intervals showed underestimation in capturing the high flows, while slight overestimation at low to intermediate and high to low flow conditions was also observed. This might be owing to inherited model biases that persisted due to too few training samples for the extremes (low and high flows) as compared to the intermediate values. This also has to do with the nature of the data, as generally normal perennial flow conditions are in the range of normal distribution. This implies that the normal flow fluctuates near the mean value of data every year. As these values occur more frequently, most of the data comprise these types of sample points. On the other hand, outliers in the data represent the extreme flow events that defies mean as the central tendency and hence are less frequent with fewer data points. Thus, the models possess a lesser tendency to predict outliers as compared to data lying in a normally distributed range. Alternatively, it could be due to errors in discharge measurement in floods due to extrapolation of rating curves, etc.
Figure 3

Observation and forecasting interval of FDC analysis.

Figure 3

Observation and forecasting interval of FDC analysis.

Close modal

In this study, a hybrid model of Wavelet Transforms and ANN (WT–ANN) was tested for forecasting flows at short (T0.33, T0.66), intermediate (T1), and long (T2, T4 , and T8) time intervals. The model was built using 10-daily flows of the Gilgit River.

The Gilgit River Basin is influenced by two major factors, i.e., the snow-melting process and local orography. While modeling its streamflow signals, the stochastic disturbances caused by these factors could add up to uncertainty in the simulation results. Our study proposed to leverage the utility of the ANN model by resorting to a different level of streamflow signal decomposition (approximation and details) using DWT. We decomposed the signal up to 10 levels using Daubechies wavelets. Further, in order to enhance the ANN processing, we first performed a cross-correlation analysis and selected highly correlated inputs. Then, from the selected inputs, nine input configurations were formed based upon two criteria: firstly, the best variables were selected considering the same interval. Secondly, the best variables were selected from other intervals, provided the interval was less than or equal to the reference interval. The simulations were evaluated using statistical metrics, i.e., R2, NSE, and RMSE. The results showed that the model performed best for short-term forecasting. At intermediate intervals, its performance declined and, interestingly, recovered back for long-term forecasting. This performance recovery might be due to the inverse transformation of the signal pattern (especially after T2), as indicated by the incrementing negative correlation in Table 2. This can be attributed to the snow-melting and accumulation process that plays a pivotal role in manipulating the streamflow signal in the Gilgit Basin and hence signifies the role of the annual recurrent pattern. It is also to be noted that research has exclusively considered flow as a primary input for modeling, which itself is the product of climate variables (precipitation and temperature fluxes). Therefore, the ingested role of these variables also played a part in the signal transformation. Further FDC analysis showed that short-term forecasting performed well at low and intermediate flow time scales, while beyond that the signal starts to deteriorate. This might be owing to fewer training samples for the extremes (low and high flows).

Although short-term forecasting provided better results, it did not necessarily mean that long-term forecasting will yield poor results, as it depends on several factors, e.g., the time resolution of forecasting interval (daily, 10-daily, monthly, etc.) and the behavior of the stream in that time resolution (perennial, non-perennial). The typical behavior of streamflow at short time intervals follows a predictable extrapolated trend (continuation of previous series) and hence yields good results. Depending upon behaviors and type of stream, the longer intervals could also benefit from the larger recurrent pattern of the stream (in our case perennial annual cycle regulated by snow-melting) and can improve its inference (in the case of a non-stochastic, perennial stream). However, forecasting at intermediate intervals could be challenging as they lack in capacity to capture both predictable patterns and larger recurrent patterns due to interval differences. Furthermore, the analysis also revealed that the lesser number of samples related to extreme values (low and high flow) in the training set limits the predictability of extreme events and could induce biases in results, thus adversely affecting the forecasting model.

While this study, adopted a suitable methodology, justified by peer-reviewed literature, limitations persist that must be addressed. It is also important to emphasize that although ANNs provide ease in utility, they still require an overabundance of data to train appropriately, while physically based models have no such restriction. Thereby, a physically based model may be considered a more suitable choice in low data constraints. Another challenge is non-stationarity in hydroclimatology, which has become more pronounced after the emergence of climate change. ANN is weak in predicting values that are out of bound to their training samples; therefore, the modeler must be vigilant to train a model on sufficiently longer records and use variables that coincide with extreme values of flow. Moreover, as described earlier in the conclusion that streamflow ingests the effects of climate variables, which can be reflected in predicted values, thereby validity of forecast models maybe also be subject to the intensity of climate change.

S.H., S.A, and K.Z.J. conceptualized the study; Z.S., RF., and K.A. did data curation; Z.S and P.M. did formal analysis; Z.S. and P.M. investigated the study; S.A., K.Z.J., and Z.S. performed the methodology; S.H., S.A., and S.S. collected resources; S.H. and Z.S wrote the original draft.

This research received no external funding.

The authors have no relevant financial or non-financial interests to disclose.

Data cannot be made publicly available; readers should contact the corresponding author for details.

The authors declare there is no conflict.

Adamowski
J.
&
Chan
H.
2011
A wavelet neural network conjunction model for groundwater level forecasting
.
Journal of Hydrology
407
,
28
40
.
https://doi.org/10.1016/j.jhydrol.2011.06.013
.
Adamowski
J.
&
Sun
K.
2010
Development of a coupled wavelet transform and neural network method for flow forecasting of non-perennial rivers in semi-arid watersheds
.
Journal of Hydrology
390
,
85
91
.
https://doi.org/10.1016/j.jhydrol.2010.06.033
.
Adnan
M.
,
Nabi
G.
,
Kang
S.
,
Zhang
G.
,
Anjum
M. N.
,
Adnan
R.
,
Iqbal
M.
&
Ali
A.
2016a
Snowmelt runoff modelling under projected climate change patterns in the Gilgit River Basin of Northern Pakistan
.
Polish Journal of Environmental Studies
26
.
https://doi.org/10.15244/pjoes/66719
.
Adnan
M.
,
Nabi
G.
,
Pomee
M.
&
Ashraf
A.
2016b
Snowmelt runoff prediction under changing climate in the Himalayan Cryosphere: a case of Gilgit River Basin
.
Geoscience Frontiers
8
.
https://doi.org/10.1016/j.gsf.2016.08.008
.
Akhtar
M.
,
Corzo
G.
,
Andel
S.
&
Jonoski
A.
2009
River flow forecasting with artificial neural networks using satellite observed precipitation pre-processed with flow length and travel time information: case study of the Ganges River Basin
.
Hydrology and Earth System Sciences Discussions
6
.
https://doi.org/10.5194/hess-13-1607-2009
.
Ali
S.
&
Shahbaz
M.
2020
Streamflow forecasting by modeling the rainfall–streamflow relationship using artificial neural networks
.
Modeling Earth Systems and Environment
6
(
3
),
1645
1656
.
https://doi.org/10.1007/s40808-020-00780-3
.
Apaydin
H.
,
Feizi
H.
,
Sattari
M.
,
Çolak
M. S.
,
Band
S.
&
Chau
K.
2020
Comparative analysis of recurrent neural network architectures for reservoir inflow forecasting
.
Water
.
https://doi.org/10.3390/w12051500
.
Aqil
M.
,
Kita
I.
,
Yano
A.
&
Nishiyama
S.
2007
Neural networks for real time catchment flow modeling and prediction
.
Water Resources Management
21
,
1781
1796
.
https://doi.org/10.1007/s11269-006-9127-y
.
Bashir
F.
&
Rasul
G.
2010
Estimation of water discharge from Gilgit basin using remote sensing, GIS and runoff modeling
.
Pakistan Journal of Meteorology
6
.
Cogley
J.
2010
A more complete version of the world glacier inventory
.
Annals of Glaciology
50
.
https://doi.org/10.3189/172756410790595859
.
Daubechies
I.
1990
The wavelet transform, time-frequency localization and signal analysis
.
Information Theory, IEEE Transactions On
36
,
961
1005
.
https://doi.org/10.1109/18.57199
.
Freire
P.
,
Santos
C.
&
Silva
G.
2019
Analysis of the use of discrete wavelet transforms coupled with ANN for short-term streamflow forecasting
.
Applied Soft Computing
81
.
https://doi.org/10.1016/j.asoc.2019.04.024
.
Govindaraju
R.
2000
Artificial neural networks in hydrology. II: hydrologic applications
.
Journal of Hydrologic Engineering
5
,
124
137
.
https://doi.org/10.1061/(ASCE)1084-0699(2000)5:2(124)
.
Hassan
Z.
,
Shamsudin
S.
&
Harun
S.
2014
Minimum input variances for modelling rainfall-runoff using ANN
.
Jurnal Teknologi (Sciences and Engineering)
69
,
113
118
.
https://doi.org/10.11113/jt.v69.3154
.
Hayder
G.
,
Solihin
M. I.
&
Najwa
M.
2022
Multi-step-ahead prediction of river flow using NARX neural networks and deep learning LSTM
.
H2Open Journal
5
.
https://doi.org/10.2166/h2oj.2022.134
.
Honorato
A.
,
Silva
G.
&
Santos
C.
2019
Monthly streamflow forecasting using neuro-wavelet techniques and input analysis
.
Hydrological Sciences Journal
63
,
1
16
.
https://doi.org/10.1080/02626667.2018.1552788
.
Hosseini
H.
,
Hashemi
H.
,
Fakheri Fard
A.
&
Berndtsson
R.
2022
Areal precipitation coverage ratio for enhanced AI modelling of monthly runoff: a new satellite data-driven scheme for semi-arid mountainous climate
.
Remote Sensing
14
,
270
.
https://doi.org/10.3390/rs14020270
.
Islam
Z.
2011
A Review on Physically Based Hydrologic Modeling
.
https://doi.org/10.13140/2.1.4544.5924
.
Jianjin
W.
,
Shi
P.
,
Jiang
P.
,
Hu
J.
,
Qu
S.
,
Chen
X.
,
Chen
Y.
,
Dai
Y.
&
Xiao
Z.
2017
Application of BP neural network algorithm in traditional hydrological model for flood forecasting
.
Water
9
,
48
.
https://doi.org/10.3390/w9010048
.
Jimeno-Sáez
P.
,
Senent-Aparicio
J.
,
Pérez-Sánchez
J.
&
Pulido-velazquez
D.
2018
A comparison of SWAT and ANN models for daily runoff simulation in different climatic zones of peninsular Spain
.
Water
10
.
https://doi.org/10.3390/w10020192
.
Khan
M. A.
,
Stamm
J.
&
Haider
S.
2021
Assessment of soft computing techniques for the prediction of suspended sediment loads in rivers
.
Applied Sciences
11
,
8290
.
https://doi.org/10.3390/app11188290
.
Kisi
O.
&
Uncuoğlu
E.
2005
Comparison of three back-propagation training algorithms for two case studies
.
Indian Journal of Engineering and Materials Sciences
12
.
Latif
Y.
,
Ma
Y.
,
Ma
W.
,
Muhammad
S.
&
Yaseen
M.
2019
Snowmelt Runoff Simulation During Early 21st Century Using Hydrological Modelling in the Snow-Fed Terrain of Gilgit River Basin (Pakistan)
. pp.
73
76
.
https://doi.org/10.1007/978-3-030-01572-5_18
.
Nacar
S.
,
Hınıs
M.
&
Kankal
M.
2017
Forecasting daily streamflow discharges using various neural network models and training algorithms
.
KSCE Journal of Civil Engineering
.
https://doi.org/10.1007/s12205-017-1933-7
.
Noori
N.
&
Kalin
L.
2016
Coupling SWAT and ANN models for enhanced daily streamflow prediction
.
Journal of Hydrology
533
,
141
151
.
https://doi.org/10.1016/j.jhydrol.2015.11.050
.
Partal
T.
&
Cigizoglu
H.
2008
Estimation and forecasting of daily suspended sediment data using wavelet-neural networks
.
Journal of Hydrology
358
,
317
331
.
https://doi.org/10.1016/j.jhydrol.2008.06.013
.
Pitta
S.
,
Venkata Praveen
T.
&
Prasad
M.
2016
Artificial neural network model for rainfall-runoff – a case study
.
International Journal of Hybrid Information Technology
9
,
263
272
.
Qasem
S.
,
Samadianfard
S.
,
Kheshtgar
S.
,
Jarhan
S.
,
Kisi
O.
,
Band
S.
&
Chau
K.
2019
Modeling monthly pan evaporation using wavelet support vector regression and wavelet artificial neural networks in arid and humid climates
.
Engineering Applications of Computational Fluid Mechanics
13
,
177
187
.
https://doi.org/10.1080/19942060.2018.1564702
.
Santos
C.
&
Morais
B.
2013
Identification of precipitation zones within São Francisco River Basin (Brazil) by global wavelet power spectra
.
Hydrological Sciences Journal
58
,
789
796
.
https://doi.org/10.1080/02626667.2013.778412
.
Santos
C.
&
Silva
G.
2014
Daily streamflow forecasting using a wavelet transform and artificial neural network hybrid models
.
Hydrological Sciences Journal
59
,
312
324
.
https://doi.org/10.1080/02626667.2013.800944
.
Santos
C.
,
Freire
P.
,
Akrami
S.
&
da Silva
R.
2019
Hybrid wavelet neural network approach for daily inflow forecasting using tropical rainfall measuring mission data
.
Journal of Hydrologic Engineering
24
.
https://doi.org/10.1061/(ASCE)HE.1943-5584.0001725
.
Šaur
D.
2015
Evaluation of the accuracy of numerical weather prediction models
.
Advances in Intelligent Systems and Computing
347
,
181
191
.
https://doi.org/10.1007/978-3-319-18476-0_19
.
Schreiner-Mcgraw
A.
&
Ajami
H.
2020
Impact of uncertainty in precipitation forcing data sets on the hydrologic budget of an integrated hydrologic model in Mountainous Terrain
.
Water Resources Research
56
.
https://doi.org/10.1029/2020WR027639
.
Scijinks
n.d.
How Reliable Are Weather Forecasts?
.
Senthil Kumar
A.
,
Goyal
M.
,
Ojha
C.
,
Singh
R.
&
Swamee
P.
2013
Application of artificial neural network, fuzzy logic and decision tree algorithms for modelling of streamflow at Kasol in India
.
Water Science and Technology : A Journal of the International Association on Water Pollution Research
68
,
2521
2526
.
https://doi.org/10.2166/wst.2013.491
.
Seydou
T.
,
Wang
Y.-M.
&
Kerh
T.
2010
Artificial neural network for modeling reference evapotranspiration complex process in Sudano-Sahelian zone
.
Agricultural Water Management
97
,
707
714
.
https://doi.org/10.1016/j.agwat.2010.01.002
.
Syed
Z.
,
Ahmad
S.
,
Dahri
Z.
,
Azmat
M.
,
Shoaib
M.
,
Inam
A.
,
Qamar
M.
,
Hussain
S.
&
Ahmad
S.
2022
Hydroclimatology of the Chitral River in the Indus Basin under changing climate
.
Atmosphere
13
,
295
.
https://doi.org/10.3390/atmos13020295
.
Tabbussum
R.
&
Qayoom
D.
2020
Comparative analysis of neural network training algorithms for the flood forecast modelling of an alluvial Himalayan River
.
Journal of Flood Risk Management
13
.
https://doi.org/10.1111/jfr3.12656
.
Wang
K.
,
Band
S. S.
,
Ameri
R.
,
Biyari
M.
,
Hai
T.
,
Hsu
C.-C.
,
Hadjouni
M.
,
Elmannai
H.
,
Chau
K.
&
Mosavi
A.
2022
Performance improvement of machine learning models via wavelet theory in estimating monthly river streamflow
.
Engineering Applications of Computational Fluid Mechanics
16
,
1833
1848
.
https://doi.org/10.1080/19942060.2022.2119281
.
Wunsch
A.
,
Liesch
T.
&
Broda
S.
2020
Groundwater level forecasting with artificial neural networks: a comparison of LSTM, CNN and NARX
.
https://doi.org/10.5194/hess-2020-552
.
Xu
B.
,
Wang
N.
,
Chen
T.
&
Li
M.
2015
Empirical Evaluation of Rectified Activations in Convolutional Network
.

Author notes

These authors contributed equally to this work.

This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (CC BY 4.0), which permits copying, adaptation and redistribution, provided the original work is properly cited (http://creativecommons.org/licenses/by/4.0/).

Supplementary data