Accurate prediction of monthly runoff is critical for effective water resource management and flood forecasting in river basins. In this study, we developed a hybrid deep learning (DL) model, Fourier transform long short-term memory (FT-LSTM), to improve the prediction accuracy of monthly discharge time series in the Brahmani river basin at Jenapur station. We compare the performance of FT-LSTM with three popular DL models: LSTM, recurrent neutral network, and gated recurrent unit, considering different lag periods (1, 3, 6, and 12). The lag period, representing the interval between the observed data points and the predicted data points, is crucial for capturing the temporal relationships and identifying patterns within the hydrological data. The results of this study show that the FT-LSTM model consistently outperforms other models across all lag periods in terms of error metrics. Furthermore, the FT-LSTM model demonstrates higher Nash–Sutcliffe efficiency and R2 values, indicating a better fit between predicted and actual runoff values. This work contributes to the growing field of hybrid DL models for hydrological forecasting. The FT-LSTM model proves effective in improving the accuracy of monthly runoff forecasts and offers a promising solution for water resource management and river basin decision-making processes.

Runoff modeling plays a vital role in hydrology and water resource management. Runoff refers to the portion of rainfall or snowmelt that flows over the land surface and eventually enters streams, rivers, and other water bodies. Precise prediction of daily runoff is essential for various applications, including flood forecasting, water supply planning, irrigation management, and environmental impact assessments (Halwatura & Najim 2013; Liu et al. 2020). The accurate prediction of monthly runoff in river basins holds paramount importance for effective water resource management, flood forecasting, and sustainable development initiatives. In the context of the Brahmani river basin at Jenapur station, India, accurate forecasting plays a pivotal role in enabling informed decision-making processes for sustainable water resource allocation and management. Traditionally, hydrologists have relied on empirical and physical models to estimate runoff (Scharffenberg et al. 2010). These models often require extensive data inputs, such as precipitation, temperature, soil properties, and land cover information (Legesse et al. 2003). While these models have been successful to some extent, they often face challenges in capturing the complex nonlinear relationships and spatial-temporal dynamics inherent in hydrological processes (Gain et al. 2021). The existing literature has emphasized the challenges associated with achieving precise runoff predictions, underscoring the necessity for advanced forecasting techniques that can accommodate the complex interplay of hydrological processes. While traditional methods have yielded valuable insights, their limitations in capturing intricate nonlinear relationships within the hydrological system have prompted the exploration of more sophisticated modeling approaches. In the past few years, deep learning (DL) models have gained significant prominence as a robust tool for analyzing intricate data patterns and making precise predictions across diverse fields (Miotto et al. 2018). DL, a subfield of machine learning, involves training artificial neural networks (ANNs) with multiple hidden layers to learn hierarchical representations of data (Ongsulee 2017). These techniques have shown remarkable success in computer vision, natural language processing, and speech recognition (Torfi et al. 2020). Applying DL techniques to daily runoff modeling holds great promise for improving prediction accuracy and capturing intricate hydrological processes (Ji et al. 2021). DL models can automatically learn features and representations from raw data, enabling them to effectively handle nonlinear relationships, temporal dependencies, and spatial patterns that influence runoff (Hu et al. 2018). Moreover, DL techniques have the potential to leverage the vast amounts of hydrological data available today, including remotely sensed data, weather station records, and historical runoff measurements (Zhu et al. 2023).

Traditional hydrological modeling approaches have been widely used for predicting and understanding daily runoff (Mao et al. 2021). These models are typically based on empirical or physical methods that aim to simulate the hydrological processes and relationships governing runoff generation (Legesse et al. 2003). Some commonly used traditional modeling approaches are discussed below.

Empirical models: These models are constructed on statistical relationships between observed runoff and meteorological variables (Thornton et al. 1997). They assume that historical patterns and correlations can be used to predict future runoff. Examples include regression models, autoregressive models, and ANNs with a simplified architecture (Abbott et al. 1986; Furundzic 1998; Jain & Kumar 2007).

Conceptual models: These models are built on simplified representations of hydrological processes (Atkinson et al. 2012). They divide a catchment into hydrological components, such as rainfall excess, soil moisture, and groundwater storage (Skøien et al. 2003). These models utilize conceptual equations and parameters to simulate the flow and storage dynamics. Examples include the soil conservation service-curve number method and the variable infiltration capacity model (Kheimi & Abdelaziz 2022).

Distributed physical models: These models simulate the hydrological processes at a spatially distributed level within a catchment. They account for spatial variations in topography, soil properties, land use, and meteorological inputs (Abbott et al. 1986). These models typically employ mathematical equations representing the physical principles of hydrological processes, such as rainfall–runoff transformation, infiltration, evapotranspiration, and flow routing (Ramírez 2000). Examples include the soil and water assessment tool, the hydrological modeling system, and the MIKE Système Hydrologique Européen (MIKE-SHE) models.

Rainfall–runoff models: These models focus on the transformation of rainfall into runoff without explicitly considering the detailed physical processes (Jakeman & Hornberger 1993). These models often use lumped or semidistributed approaches and rely on calibration against observed runoff data (Ruelland et al. 2008). Examples include the tank model, the rational method, and the unit hydrograph method (Suryoputro et al. 2017).

Statistical time series models: These models focus on analyzing the statistical properties of hydrological variables, such as rainfall and runoff (Efstratiadis et al. 2014). They utilize techniques like autoregressive integrated moving average, seasonal decomposition of time series, and Fourier analysis to capture temporal patterns and forecast future runoff (Zhou et al. 2021).

These traditional hydrological modeling approaches have been widely used and have contributed significantly to our understanding of runoff processes. However, they often face limitations in capturing the complex spatial-temporal dynamics and nonlinear relationships present in hydrological systems (Krause et al. 2015). DL techniques provide an alternative and potentially more accurate approach by leveraging the power of neural networks to learn intricate patterns and relationships directly from data (Kautz et al. 2017; Berisha et al. 2019).

DL techniques have revolutionized various fields by enabling machines to learn and make predictions from complex data patterns (Rajula et al. 2020). DL is a subfield of machine learning that utilizes ANNs with multiple layers to automatically learn hierarchical representations of data (Xie et al. 2017). These techniques have shown immense potential in domains such as computer vision, natural language processing, and speech recognition (Hirschberg & Manning 2015). In recent years, their application to hydrology and runoff modeling has gained attention, offering new avenues for improved prediction accuracy and understanding of hydrological processes (ASCE Task Committee on Application of Artificial Neural Networks in Hydrology 2000; Shen et al. 2018).

The potential of DL techniques in runoff modeling stems from their ability to effectively capture intricate spatial-temporal patterns, handle nonlinearity, and extract features directly from raw data (Tahmasebi et al. 2020; Ha et al. 2021). Unlike traditional hydrological models that rely on predefined equations and assumptions, DL models can learn and adapt to the data, allowing them to capture complex relationships that may be challenging to formulate explicitly (Reichstein et al. 2019).

The advantages of DL techniques for runoff modeling include the following:

Nonlinear modeling: DL models excel at capturing nonlinear relationships between input variables and runoff (Ditthakit et al. 2023). They can learn complex patterns and interactions, even in the presence of high-dimensional data (Advani et al. 2013).

Temporal dependency: DL models can effectively model temporal dependencies, considering the sequential nature of runoff data. They can capture short-term and long-term dependencies, enabling accurate predictions of runoff dynamics over time (Li et al. 2021b).

Spatial patterns: DL models have the potential to capture spatial patterns by incorporating spatially distributed data, such as rainfall, soil properties, and topography (Zandi et al. 2022). This enables them to account for the heterogeneity within a catchment and improve the accuracy of runoff predictions (Santikari & Murdoch 2018).

Data integration: DL techniques can integrate various data sources, including remotely sensed data, meteorological observations, and historical runoff measurements. This integration enhances the representation of hydrological processes and improves the model's generalization capabilities (Goetz et al. 2011).

Feature learning: DL models can learn relevant features from raw input data (Yuan et al. 2019). They can extract meaningful representations, reducing the need for manual feature engineering and potentially capturing important but previously unknown relationships in runoff processes (Kim et al. 2021).

Scalability: DL models can handle large datasets efficiently and scale well with increasing data volume (Al-Jarrah et al. 2015). This capability is particularly valuable in hydrology, where vast amounts of data are available from weather stations, remote sensing platforms, and other sources. By leveraging these advantages, DL techniques offer new opportunities for improved runoff modeling, including more accurate predictions, enhanced understanding of underlying procedures, and better management of water resources (Tahmasebi et al. 2020).

In this study, we introduce a novel hybrid DL model, Fourier transform long short-term memory (FT-LSTM), designed to enhance the prediction accuracy of monthly discharge time series. By integrating the Fourier transform with LSTM, our model aims to capture both temporal dependencies and periodic patterns present in the hydrological data, thus improving the overall forecasting performance. The Brahmani River basin at the Jenapur station in India is selected as the typical study site. The primary objective of this research is to assess the effectiveness of the FT-LSTM model in comparison to traditional DL models, including LSTM, recurrent neutral network (RNN), and gated recurrent unit (GRU) across various lag periods. In this paper, we present a comprehensive analysis of the performance of the FT-LSTM model, demonstrating its efficacy in achieving enhanced forecasting accuracy and providing valuable insights into water resource management and hydrological applications. By addressing the existing research gaps and leveraging the potential of advanced DL techniques, this study contributes to the growing body of knowledge in the field of hybrid DL models for hydrological forecasting.

The Brahmani River Basin is situated in the eastern part of India. It is the second largest river of Odisha. Its land use is a combination of forest and agriculture. Agricultural land accounts for around 52% of the basin. In the rest of the basin, forests are the major land use. The South Koel and Sankh Rivers combine to form the Brahmani river basin. Its latitude extends from 20°28′ to 23°35′ N and longitude extends from 83°52′ to 87°03′E (Figure 1). The total catchment area is found to be 39,313 km2 and receives almost about 1,305 mm of average annual rainfall. It has a tropical climate. The maximum temperature it reaches is as high as 47 °C during summer and drops to a minimum temperature of 4 °C in winter. Around 70% of the basin has a gentle slope. The maximum elevation of the basin is nearly 1,181 m.
Figure 1

Location of study area and station.

Figure 1

Location of study area and station.

Close modal

Data collection and data preprocessing

The daily runoff data from January 1990 to December 2020 of the Jenapur station were obtained from the Central Water Commission (CWC), Bhubaneswar. Guo et al. (2011) found that various elements, such as temperature, evaporation, and rainfall, have an impact on the creation process of streamflow. As rainfall, a natural phenomenon with significant temporal and spatial variability, is the primary determinant of streamflow, this research focuses on creating a streamflow forecasting model based on rainfall.

Furthermore, the input scenarios were created to explore the sensitivity by using lagged data conditions, considering the historical data as a reference for the next steps. As a result, it will be utilized to evaluate the execution of the effects and outcomes represented in the model. The number of time lags of the streamflow was determined by using the partial autocorrelation function (PACF) shown in Figure 2. The streamflow at t = 1 month was affected by the events of previous month values. To choose the model characteristics, three lag days of rainfall were taken into account. Given a series of rainfall observations (, where N is the length of time lag), the streamflow forecast can be given as
(1)
where represents the predictive streamflow at time (t+ 1), R denotes rainfall dataset, and is a forecasting model. The duration of the lagged inputs M in Equation (1) must be chosen carefully if good prediction accuracy is to be achieved.
Figure 2

Monthly time series date with ACF and PACF plot.

Figure 2

Monthly time series date with ACF and PACF plot.

Close modal

To prevent the model from overfitting and compare its performance in predicting streamflow, it is therefore clear and quick to divide the input data into training (train LSTM), validation (assess LSTM), and testing (confirm the findings).

The back values that will be used for testing and training must be taken into account when using time series data; therefore, the data were divided into three groups: training (70%), validation (15%), and testing (15%). From 1 January 1990 to 31 December 2010, the learning period was in effect. From 1 January 2010 to 31 December 2015, the validation period was in effect. From 1 January 2015 to 31 December 2020, the testing period was in effect.

Fourier transformation

FT is a mathematical technique used to transform a time domain signal into its frequency domain representation. The Fourier smoothing filter is one that can be applied to achieve a smoothing effect on a time series. The Fourier smoothing filter involves performing the FT on a time series, attenuating or removing high-frequency components (i.e., noise or rapid changes), and then applying the inverse Fourier transformation to obtain the smoothed time series. It is important to note that the Fourier smoothing filter is a linear filter and assumes that the noise or unwanted variations in the time series are represented by high-frequency components. Therefore, the effectiveness of Fourier smoothing depends on the characteristics of the specific time series and the desired smoothing effect.

In our study, the application of the Fourier Transform methodology played a crucial role in enhancing the predictive capabilities of the proposed FT-LSTM model. Fourier transform was initially employed to decompose the time series data into its frequency components. Subsequently, the low-pass filter was utilized to retain the lower-frequency components, which are indicative of the significant temporal trends and variations, while minimizing the influence of higher-frequency noise and fluctuations. By effectively harnessing the Fourier transform methodology in conjunction with the low-pass filtering technique, the FT-LSTM model was able to capitalize on the inherent temporal patterns within the monthly discharge time series, thereby enhancing the accuracy and robustness of the predictive capabilities for hydrological forecasting in the Brahmani River at the Jenapur station.

The Fourier Transform of time series signal is given:
(2)
where represents the frequency domain representation of the signal, is the time domain signal, f is the frequency, and i is the imaginary unit.

More details on FT for time series smoothing can be found in the research paper by Kimball (1974).

Figure 3 summaries the workflow of the study methodology.
Figure 3

Study methodology workflow.

Figure 3

Study methodology workflow.

Close modal

Long short-term memory

LSTM was developed by Hochreiter & Schmidhuber (1997). It is a subtype of RNNs that is utilized by a large number of researchers due to its unique architecture that incapacitates the RNNs' long-term dependency issue (Piccialli et al. 2021). It also reduces the impact of the vanishing gradient issues on training performance (Kim et al. 2016). LSTM neural network nodes receive the latest states from the preceding step, similar to RNNs. Nonetheless, the node, which is a common LSTM unit, has a more sophisticated structure than it does in an RNN, and this is the key element that provides long-term memory by reducing the vanishing gradient outcome (Kilinc 2022; Wang et al. 2022). The configuration of an LSTM is governed by three essential elements: the cell state, representing the current long-term memory of the network; the hidden state, which is the output from the previous time step; and the input data for the current time step (Gao et al. 2020; Li et al. 2021a). This LSTM architecture enables the control of information flow through three specific gates: the forget gate, the input gate, and the output gate, as illustrated in Figure 4.
Figure 4

The interior design of an LSTM cell (Park & Kim 2023).

Figure 4

The interior design of an LSTM cell (Park & Kim 2023).

Close modal
The gates in an LSTM model play a crucial role in managing internal processes, facilitating the handling of data sequences through the cell state, and coordinating the flow of information within each cell. The first internal mechanism involves the forget gate , which employs a sigmoid layer to determine which information is irrelevant and should be discarded. On the other hand, the input gate leverages both a sigmoid layer and a tanh layer to select input values for updating the memory state generating a vector of candidate values. The old cell state is then updated to a new state using this input. Furthermore, the output gate, controlled by the input values and memory block, regulates the output values. The forget gate determines what information should be ignored from the memory block. The LSTM internal processes are governed by mathematical Equations (3)–(9), as depicted below:
(3)
(4)
(5)
(6)
(7)
(8)
(9)

Recurrent neural networks

The achievements witnessed in recent times regarding the application of DL have been particularly notable in sequential prediction tasks. These tasks include statistical language modeling, analysis of chaotic time series, ecological modeling for dynamic systems control, and applications in finance and marketing. The success of DL in these domains has inspired researchers to explore its potential for time series forecasting in hydrology events (Duriez et al. 2017; Chattopadhyay et al. 2020; Pan et al. 2022)

The capability of RNNs to capture dependencies between previous terms is crucial for accurate weather pattern predictions (Banerjee et al. 2022). RNNs are particularly valuable in this context as they are ANNs specifically designed for building prediction models using long-term time series datasets (Kim et al. 2018; Chimmula & Zhang 2020). Depending on the specific application, RNN's architecture offers various variations, including the many-to-one model (used to predict the current time step based on all previous inputs), the many-to-many model (employed to predict multiple future time steps simultaneously based on all previous inputs), and other models (Sahoo et al. 2019; Bose et al. 2022). The selection of the final structure is determined by the formulation of the problem, which is influenced by the observed phenomena.

RNNs are often considered an extension of conventional feedforward neural networks, but with an added advantage of cyclic connections, which endow them with the capability to effectively model sequential data. In the context of RNNs, we assume the presence of an input sequence denoted by X, a hidden vector sequence denoted by H, and an output vector sequence denoted by Y. The input sequence is represented as . A traditional RNN calculates the hidden vector sequence and output vector sequence . In a typical RNN (Figure 5), the calculation of the hidden vector sequence and output vector sequence is performed for each time step (t = 1 to T) according to the following equations:
(10)
(11)
where function σ is a nonlinearity function, W is a weight matrix, and b is a bias term.
Figure 6

A GRU cell.

Figure 7

Scatter plot between observed vs estimated runoff during training and testing: (a) FT-LSTM; (b) GRU; (c) LSTM; and (d) RNN.

Figure 7

Scatter plot between observed vs estimated runoff during training and testing: (a) FT-LSTM; (b) GRU; (c) LSTM; and (d) RNN.

Close modal
Figure 8

Boxplots of observed and estimated monthly runoff volumes of different models.

Figure 8

Boxplots of observed and estimated monthly runoff volumes of different models.

Close modal

Gated recurrent units

The GRU is a type of RNN architecture that has gained significant popularity in the field of DL, as shown in Figure 6. It was introduced by Cho et al. (2014) as an alternative to the traditional LSTM units. GRUs are designed to address some of the limitations of LSTMs while achieving comparable performance in many tasks. Like LSTMs, GRUs are capable of capturing long-range dependencies in sequential data, making them well-suited for tasks such as speech recognition, natural language processing, and machine translation. One of the key advantages of GRUs over LSTMs is their simpler architecture. While LSTMs have separate input and output gates and a forget gate, GRUs combine the functionality of these gates into two gates: an update gate and a reset gate. This results in a more streamlined structure with fewer parameters, making GRUs computationally more efficient and easier to train. The update gate in a GRU determines how much information from the previous time step should be passed to the current time step. It controls the trade-off between incorporating new input and retaining past information. The reset gate, on the other hand, decides how much of the past information should be forgotten.

The mathematical equations that govern the behavior of a GRU can be summarized as follows:
(12)
(13)
(14)
(15)

The GRU's update gate allows it to dynamically adapt to the input and the context, controlling the flow of information through the network. This adaptability is beneficial when dealing with long sequences where retaining relevant information over long periods is crucial. Due to their simpler architecture, GRUs are generally faster to train and require fewer computational resources compared with LSTMs. However, the performance of GRUs and LSTMs can vary depending on the specific task and dataset. It is often recommended to experiment with both architectures to determine which one works best for a given problem. The GRU is a recurrent neural network architecture that addresses some of the complexities of LSTMs while achieving comparable performance. With its simplified design and efficient computational properties, the GRU has become a popular choice for modeling sequential data in various domains of DL.

Numerous RNN and LSTM models for runoff forecasting have been constructed using sequential raw data. The LSTM layer is a standard component in many recent machine learning software packages, including TensorFlow and Keras (Xiang et al. 2020). Keras is a Python-based high-level neural network library that can run on top of Theano or TensorFlow. TensorFlow is a well-known machine learning package that is commonly used for constructing DL models. Nonetheless, TensorFlow has several limits. Keras, on the other hand, is a user-friendly and easily available high-level application programming interface (API) built on top of TensorFlow (Tang 2016).

To build this model, the monthly runoff data were taken. The datasets were categorized into three sets: training, validation, and testing sets. Researchers have used diverse datasets to generate training and testing sets in various research investigations. In this study, 70% of the available dataset was allotted for training sets, with the remaining 30% allocated for validation and testing sets (15% each). The model's input generally consists of different time lags of inputs and their combinations that are shown in Table 1. Five cases of input variable (X1, X3, X6, X9, and X12) were designed with lags ranging from 1 to 12 months before the current time for runoff forecasting. It was a 3D array with the dimensions of number of samples (372); number of time steps (lag) with values of 1, 3, 6, and 12; and an output dimension of 1. In all cases, the model was trained using 70% of the sample data. The model was compiled using the mean square error loss function and the Adam optimizer, and this was the final phase in the model's construction. The model was trained for 500 epochs. A scatter plot between observed vs estimated runoff during training and testing for all the models is shown in (Figure 7).

Table 1

The optimal input value for the Brahmani River Basin

  
  
  
  
  
  
  
  
  
  

, the predicted runoff; , the 1-month lagged; , the 3-month lagged; , the 6-month lagged; , the 9-month lagged; , the 12-month lagged; f, the model type.

Input selection is considered to be the most important factor for model development. Hence, various machine learning methods were adopted for determining the input parameters, including the parsimonious method, the autocorrelation function (ACF), and the PACF (Figure 2) (Abdel-Aal & Al-Garni 1997; Kumar et al. 2019).

This study evaluates the performance of each model by employing four statistical metrics: Nash–Sutcliffe efficiency (NSE), mean absolute error (MAE), coefficient of determination (R2), and root mean squared error (RMSE). The R2 is a metric that quantifies the extent to which the model can account for the observed variance, with values ranging from 0 to 1. A value of 0 indicates no correlation, while a value of 1 indicates that the model can accurately capture all observed variances. RMSE measures the proximity between the model's predictions and the actual observations. RMSE values range from 0 (indicating a perfect fit) to +∞ (suggesting no fit), depending on the relative range of the data. The NSE, originally introduced by Sutcliffe and Nash, is frequently employed to evaluate model performance. NSE assesses the model's ability to predict different variables relative to the mean and provides a measure of the proportion of initial variance explained by the model. NSE values range from 1 (indicating an exact match) to −∞.
(16)
(17)
(18)
(19)
where is the ith model-estimated monthly runoff, is the ith observed monthly runoff. is the average of the estimated monthly runoff data, is the average of the observed monthly runoff data, and n is the number of observations.

Choice of comparative models

LSTM was chosen due to its ability to capture long-range dependencies and retain information over extended time intervals. Its effectiveness in handling sequential data and mitigating the vanishing gradient problem makes it a popular choice for time-series forecasting tasks, including hydrological modeling. RNNs, similar to LSTM, were included as a comparative model to assess the performance of the FT-LSTM model against a more basic recurrent architecture. The comparison helps in understanding the benefits of incorporating specialized mechanisms, such as memory cells and gating units, as present in the LSTM and FT-LSTM models. GRU, known for its simplified gating mechanism compared with the LSTM, was selected to provide insights into the impact of different gating structures on the predictive accuracy of monthly runoff. Comparing GRU with LSTM and FT-LSTM aids in evaluating the contributions of more intricate memory management in capturing the temporal dynamics of the Brahmani River Basin's runoff data.

Considering these limitations and the rationale behind the selection of comparative models, a comprehensive analysis was conducted to ensure a fair and insightful comparison of the FT-LSTM model's performance with respect to traditional and other DL architectures.

Table 2 presents the error metrics for runoff prediction in the Brahmani River Basin at the Jenapur station, considering different lag periods (1, 3, 6, and 12) for four forecasting models: LSTM, RNN, GRU, and FT-LSTM. When comparing the models, it is evident that the FT-LSTM consistently outperforms the other models in terms of error metrics across all lag periods. For example, at lag 12, the FT-LSTM model achieves an MAE of 487.65 (m3/s), which is significantly lower than the MAEs of the other models: LSTM (795.57 m3/s), RNN (841.06 m3/s), and GRU (829.69 m3/s). Similarly, the RMSE for FT-LSTM at lag 12 (592.58 m3/s) is substantially lower compared with LSTM (1,028.22 m3/s), RNN (1,074.98 m3/s), and GRU (1,102.37 m3/s).

Table 2

The summary of performance comparison for the used models applied in Jenapur station for different lags

Lag 1
Lag 3
Lag 6
Lag 12
TrainingValidateTestingTrainingValidateTestingTrainingValidateTestingTrainingValidateTesting
LSTM MAE (m3/s) 1,423.22 1,312.4 1,197.43 1,019.18 1,000.17 859.26 871.95 1,004.15 845.66 711.59 852 795.57 
RMSE (m3/s) 2,098.42 1,887.53 1,632.68 1,604.18 1,622.03 1,143.38 1,493.97 1,654.39 1,174.15 1,273.29 1,522.74 1,028.22 
NSE 0.45 0.38 0.41 0.68 0.55 0.68 0.72 0.55 0.67 0.8 0.56 0.76 
R2 0.46 0.39 0.44 0.68 0.59 0.75 0.73 0.61 0.78 0.81 0.62 0.83 
RNN MAE (m3/s) 1,485.09 1,384.42 1,225.78 1,269.7 1,228.86 983.25 1,082.05 1,092.04 920.06 936.39 885.5 841.06 
RMSE (m3/s) 2,155.85 1,970.24 1,624.76 1,901.72 1,829.54 1,310.1 1,741.21 1,686.36 1,252.1 1,600.79 1,608.83 1,074.98 
NSE 0.42 0.32 0.41 0.55 0.63 0.58 0.63 0.55 0.53 0.69 0.51 0.74 
R2 0.42 0.33 0.42 0.56 0.44 0.59 0.63 0.55 0.65 0.7 0.55 0.76 
GRU MAE (m3/s) 1,416.38 1,310.94 1,192.48 1,040.16 1,033.89 869.51 902.87 1,004.88 822.54 723.21 816.64 829.69 
RMSE (m3/s) 2,099.73 1,906.22 1,628.77 1,642.75 1,719.03 1,229.74 1,533.33 1,661.71 1,120.77 1,282.98 1,512.33 1,102.37 
NSE 0.45 0.36 0.41 0.67 0.5 0.63 0.71 0.55 0.7 0.8 0.57 0.73 
R2 0.45 0.38 0.43 0.67 0.55 0.71 0.71 0.59 0.76 0.8 0.64 0.82 
FT-LSTM MAE (m3/s) 1,296.65 1,089.1 1,173.15 657.91 627.64 556.2 556.56 605.5 438.18 406.22 428.49 487.65 
RMSE (m3/s) 1,774.11 1,435.05 1,437.42 992.06 847.11 762.01 830.42 873.01 647.13 539.35 569.48 592.58 
NSE 0.55 0.56 0.51 0.86 0.85 0.85 0.9 0.84 0.89 0.96 0.92 0.92 
R2 0.55 0.57 0.52 0.86 0.86 0.87 0.91 0.85 0.91 0.97 0.94 0.93 
Lag 1
Lag 3
Lag 6
Lag 12
TrainingValidateTestingTrainingValidateTestingTrainingValidateTestingTrainingValidateTesting
LSTM MAE (m3/s) 1,423.22 1,312.4 1,197.43 1,019.18 1,000.17 859.26 871.95 1,004.15 845.66 711.59 852 795.57 
RMSE (m3/s) 2,098.42 1,887.53 1,632.68 1,604.18 1,622.03 1,143.38 1,493.97 1,654.39 1,174.15 1,273.29 1,522.74 1,028.22 
NSE 0.45 0.38 0.41 0.68 0.55 0.68 0.72 0.55 0.67 0.8 0.56 0.76 
R2 0.46 0.39 0.44 0.68 0.59 0.75 0.73 0.61 0.78 0.81 0.62 0.83 
RNN MAE (m3/s) 1,485.09 1,384.42 1,225.78 1,269.7 1,228.86 983.25 1,082.05 1,092.04 920.06 936.39 885.5 841.06 
RMSE (m3/s) 2,155.85 1,970.24 1,624.76 1,901.72 1,829.54 1,310.1 1,741.21 1,686.36 1,252.1 1,600.79 1,608.83 1,074.98 
NSE 0.42 0.32 0.41 0.55 0.63 0.58 0.63 0.55 0.53 0.69 0.51 0.74 
R2 0.42 0.33 0.42 0.56 0.44 0.59 0.63 0.55 0.65 0.7 0.55 0.76 
GRU MAE (m3/s) 1,416.38 1,310.94 1,192.48 1,040.16 1,033.89 869.51 902.87 1,004.88 822.54 723.21 816.64 829.69 
RMSE (m3/s) 2,099.73 1,906.22 1,628.77 1,642.75 1,719.03 1,229.74 1,533.33 1,661.71 1,120.77 1,282.98 1,512.33 1,102.37 
NSE 0.45 0.36 0.41 0.67 0.5 0.63 0.71 0.55 0.7 0.8 0.57 0.73 
R2 0.45 0.38 0.43 0.67 0.55 0.71 0.71 0.59 0.76 0.8 0.64 0.82 
FT-LSTM MAE (m3/s) 1,296.65 1,089.1 1,173.15 657.91 627.64 556.2 556.56 605.5 438.18 406.22 428.49 487.65 
RMSE (m3/s) 1,774.11 1,435.05 1,437.42 992.06 847.11 762.01 830.42 873.01 647.13 539.35 569.48 592.58 
NSE 0.55 0.56 0.51 0.86 0.85 0.85 0.9 0.84 0.89 0.96 0.92 0.92 
R2 0.55 0.57 0.52 0.86 0.86 0.87 0.91 0.85 0.91 0.97 0.94 0.93 

The superiority of the FT-LSTM model is consistent across all lag periods, as indicated by consistently lower MAE and RMSE values. In addition, the FT-LSTM model demonstrates higher NSE and R2 values, suggesting a better fit between predicted and actual runoff values. This indicates that the FT-LSTM model is better at explaining the variance in the actual runoff data. For instance, at lag 12, the FT-LSTM model achieves an NSE of 0.92 and an R2 of 0.93, which are notably higher compared with the values obtained by LSTM, RNN, and GRU. The accuracy percentage of FT-LSTM is notably higher by 10.75%, 18.27%, and 11.82% when compared with LSTM, RNN, and GRU, respectively. The runoff volumes of observed and estimated values from different models is given in Figure 8. This indicates that the FT-LSTM model provides more accurate and reliable runoff predictions. The results of FT-LSTM were also compared with other models of previous studies to provide its reliability and accuracy, shown in Table 3.

Table 3

Obtained results of the study compared with previous studies

ModelTime seriesNSE
Moon et al. (2023)  
  • XP-storm water management model (SWMM)

 
Monthly 0.82 
Ditthakit et al. (2023)  
  • Support vector regression – radial basis function (SVR-rbf)

 
Monthly 0.59 
  • Random forest (RF)

 
0.67 
  • Model tree (M5)

 
0.57 
  • Support vector regression with polynomial kernel function (SVR-poly)

 
0.47 
Alarcon et al. (2022)  
  • Convolutional neutral network (CNN)

 
Monthly 0.81 
  • Multilayer perceptron

 
0.42 
  • LSTM

 
0.38 
Yan et al. (2022)  
  • Hybrid kernel support vector machine (HKSVM)

 
Monthly 0.90 
  • Extreme learning machine (ELM)

 
0.91 
  • Generalized regression neural network (GRNN)

 
0.87 
  • Multiple linear regression (MLR)

 
0.88 
  • LSTM

 
0.87 
  • Seasonal autoregressive integrated moving average (SARIMA)

 
0.86 
Tan et al. (2022)  
  • LSTM

 
Monthly 0.90 
Adnan et al. (2021)  
  • ELM

 
Monthly 0.82 
  • Particle swarm optimization (PSO)

 
0.88 
  • Grey wolf optimization (GWO)

 
0.89 
Cheng et al. (2020)  
  • ANN

 
Daily 0.69 
  • LSTM

 
0.76 
Rezaei et al. (2019)  
  • Ensemble empirical mode decomposition (EEMD) coupled with support vector machine (SVM)

 
Monthly 0.79 
  • EEMD model tree (EEMD-MT)

 
0.89 
  • SVMs

 
0.46 
  • Model tree (MT)

 
0.65 
Mehdizadeh et al. (2019)  Ocmulgee station Monthly  
  • Fractionally autoregressive integrated moving average (FARIMA)

 
0.90 
  • Self-exciting threshold autoregressive (SETAR)

 
0.90 
Umpqua station  
  • FARIMA

 
0.89 
  • SETAR

 
0.90 
Zakhrouf et al. (2023)  Sidi Aich Station Monthly  
  • Elman recurrent neural network (ERNN)

 
0.72 
  • LSTM

 
0.71 
  • GRU

 
0.73 
  • Feedforward neural network (FFNN)

 
0.69 
Ponteba Defluent Station  
  • ERNN

 
0.84 
  • LSTM

 
0.84 
  • GRU

 
0.87 
  • FFNN

 
0.84 
Our study 
  • LSTMs

 
Monthly 0.76 
  • RNN

 
0.74 
  • GRU

 
0.73 
  • FT-LSTM

 
0.92 
ModelTime seriesNSE
Moon et al. (2023)  
  • XP-storm water management model (SWMM)

 
Monthly 0.82 
Ditthakit et al. (2023)  
  • Support vector regression – radial basis function (SVR-rbf)

 
Monthly 0.59 
  • Random forest (RF)

 
0.67 
  • Model tree (M5)

 
0.57 
  • Support vector regression with polynomial kernel function (SVR-poly)

 
0.47 
Alarcon et al. (2022)  
  • Convolutional neutral network (CNN)

 
Monthly 0.81 
  • Multilayer perceptron

 
0.42 
  • LSTM

 
0.38 
Yan et al. (2022)  
  • Hybrid kernel support vector machine (HKSVM)

 
Monthly 0.90 
  • Extreme learning machine (ELM)

 
0.91 
  • Generalized regression neural network (GRNN)

 
0.87 
  • Multiple linear regression (MLR)

 
0.88 
  • LSTM

 
0.87 
  • Seasonal autoregressive integrated moving average (SARIMA)

 
0.86 
Tan et al. (2022)  
  • LSTM

 
Monthly 0.90 
Adnan et al. (2021)  
  • ELM

 
Monthly 0.82 
  • Particle swarm optimization (PSO)

 
0.88 
  • Grey wolf optimization (GWO)

 
0.89 
Cheng et al. (2020)  
  • ANN

 
Daily 0.69 
  • LSTM

 
0.76 
Rezaei et al. (2019)  
  • Ensemble empirical mode decomposition (EEMD) coupled with support vector machine (SVM)

 
Monthly 0.79 
  • EEMD model tree (EEMD-MT)

 
0.89 
  • SVMs

 
0.46 
  • Model tree (MT)

 
0.65 
Mehdizadeh et al. (2019)  Ocmulgee station Monthly  
  • Fractionally autoregressive integrated moving average (FARIMA)

 
0.90 
  • Self-exciting threshold autoregressive (SETAR)

 
0.90 
Umpqua station  
  • FARIMA

 
0.89 
  • SETAR

 
0.90 
Zakhrouf et al. (2023)  Sidi Aich Station Monthly  
  • Elman recurrent neural network (ERNN)

 
0.72 
  • LSTM

 
0.71 
  • GRU

 
0.73 
  • Feedforward neural network (FFNN)

 
0.69 
Ponteba Defluent Station  
  • ERNN

 
0.84 
  • LSTM

 
0.84 
  • GRU

 
0.87 
  • FFNN

 
0.84 
Our study 
  • LSTMs

 
Monthly 0.76 
  • RNN

 
0.74 
  • GRU

 
0.73 
  • FT-LSTM

 
0.92 

The consistent and superior performance of the FT-LSTM model across all lag periods demonstrates its effectiveness in accurately forecasting runoff levels in the Brahmani River Basin at the Jenapur station. The FT-LSTM model's feature transformation capabilities likely contribute to its ability to capture complex patterns in the data and make more accurate predictions.

Potential limitations and challenges

Data availability and quality: Variations in data availability and quality across different models could impact the comparative analysis. Insufficient data points or inconsistencies in data quality might affect the performance assessment, particularly in cases where certain models are more sensitive to data characteristics.

Model complexity and interpretability: Comparing the FT-LSTM model with traditional models and other DL architectures might pose challenges in terms of model complexity and interpretability. Ensuring a fair comparison while considering the interpretability of results remains crucial, especially in scenarios where simpler models might be favored due to their ease of understanding and implementation.

Computational requirements: DL models often require significant computational resources, which could pose constraints on the practical implementation and scalability of the models. A thorough evaluation of the computational requirements and efficiency is essential to understand the practical feasibility of deploying these models in real-world hydrological forecasting scenarios.

The purpose of this study is to improve the prediction accuracy of monthly runoff time series in the Brahmani River Basin at the Jenapur station. To achieve this, we compared the performance of the following three DL models: LSTM, RNN, GRU, and a new hybrid model called FT-LSTM. From the results, it was found that the FT-LSTM model outperformed other models in terms of prediction accuracy (Table 1). This shows that integrating the Fourier transform into his LSTM architecture can effectively improve the predictive ability of the model for the monthly runoff time series. The superior performance of the FT-LSTM model can be attributed to its ability to capture the inherent periodic patterns and seasonality of runoff data. By using Fourier transform as a preprocessing step, the model was able to decompose the time series into frequency components, allowing LSTM components to effectively learn complex time dependencies. The results of this study suggest that the FT-LSTM model can be expected to improve the accuracy of monthly discharge predictions in the Brahmani River Basin at the Jenapur station. The model can capture both long-term dependencies and cyclical variations in runoff data, making it a valuable tool for water resource management, flood forecasting, and other hydrological applications.

All relevant data are available from an online repository or repositories: https://github.com/bibhuatGitHUb/Accurate-prediction-of-monthly-runoff.

The authors declare there is no conflict.

Adnan
R. M.
,
Mostafa
R. R.
,
Kisi
O.
,
Yaseen
Z. M.
,
Shahid
S.
&
Zounemat-Kermani
M.
2021
Improving streamflow prediction using a new hybrid ELM model combined with hybrid particle swarm optimization and grey wolf optimization
.
Knowledge-Based Systems
230
,
107379
.
Advani
M.
,
Lahiri
S.
&
Ganguli
S.
2013
Statistical mechanics of complex neural systems and high dimensional data
.
Journal of Statistical Mechanics: Theory and Experiment
2013
(
03
),
P03014
.
Alarcon
A.
,
Garcia-Cortes
D.
,
Fernández-Alvarez
J. C.
&
Martínez-González
Y.
2022
Improving monthly rainfall forecast in a watershed by combining neural networks and autoregressive models
.
Environmental Processes
9
(
3
),
53
79
.
Al-Jarrah
O. Y.
,
Yoo
P. D.
,
Muhaidat
S.
,
Karagiannidis
G. K.
&
Taha
K.
2015
Efficient machine learning for big data: A review
.
Big Data Research
2
(
3
),
87
93
.
ASCE Task Committee on Application of Artificial Neural Networks in Hydrology.
2000
Artificial neural networks in hydrology. II: Hydrologic applications
.
Journal of Hydrologic Engineering
5
(
2
),
124
137
.
Atkinson
R.
,
Dornblut
I.
&
Smith
D.
2012
An international standard conceptual model for sharing references to hydrologic features
.
Journal of Hydrology
424
,
24
36
.
Banerjee
T.
,
Sinha
S.
&
Choudhury
P.
2022
Long term and short term forecasting of horticultural produce based on the LSTM network model
.
Applied Intelligence
52
,
9117
9147
.
Berisha
S.
,
Lotfollahi
M.
,
Jahanipour
J.
,
Gurcan
I.
,
Walsh
M.
,
Bhargava
R.
,
Van Nguyen
H.
&
Mayerich
D.
2019
Deep learning for FTIR histology: Leveraging spatial and spectral features with convolutional neural networks
.
Analyst
144
(
5
),
1642
1653
.
Bose
R.
,
Pintar
A.
&
Simiu
E.
2022
A real time prediction methodology for hurricane evolution using LSTM recurrent neural networks
.
Neural Computing and Applications
34
(
20
),
17491
17505
.
Cheng
M.
,
Fang
F.
,
Kinouchi
T.
,
Navon
I. M.
&
Pain
C. C.
2020
Long lead-time daily and monthly streamflow forecasting using machine learning methods
.
Journal of Hydrology
590
,
125376
.
Chimmula
V. K. R.
&
Zhang
L.
2020
Time series forecasting of COVID-19 transmission in Canada using LSTM networks
.
Chaos, Solitons & Fractals
135
,
109864
.
Cho
K.
,
Van Merriënboer
B.
,
Gulcehre
C.
,
Bahdanau
D.
,
Bougares
F.
,
Schwenk
H.
&
Bengio
Y.
2014
Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078
.
Ditthakit
P.
,
Pinthong
S.
,
Salaeh
N.
,
Weekaew
J.
,
Tran
T. T.
&
Pham
Q. B.
2023
Comparative study of machine learning methods and GR2M model for monthly runoff prediction
.
Ain Shams Engineering Journal
14
(
4
),
101941
.
Duriez
T.
,
Brunton
S. L.
&
Noack
B. R.
2017
Machine Learning Control-Taming Nonlinear Dynamics and Turbulence
.
Springer, Cham
.
Efstratiadis
A.
,
Dialynas
Y. G.
,
Kozanis
S.
&
Koutsoyiannis
D.
2014
A multivariate stochastic model for the generation of synthetic time series at multiple time scales reproducing long-term persistence
.
Environmental Modelling & Software
62
,
139
152
.
Gain
A. K.
,
Hossain
S.
,
Benson
D.
,
Di Baldassarre
G.
,
Giupponi
C.
&
Huq
N.
2021
Social-ecological system approaches for water resources management
.
International Journal of Sustainable Development & World Ecology
28
(
2
),
109
124
.
Gao
S.
,
Huang
Y.
,
Zhang
S.
,
Han
J.
,
Wang
G.
,
Zhang
M.
&
Lin
Q.
2020
Short-term runoff prediction with GRU and LSTM networks without requiring time step optimization during sample generation
.
Journal of Hydrology
589
,
125188
.
Goetz
J. N.
,
Guthrie
R. H.
&
Brenning
A.
2011
Integrating physical and empirical landslide susceptibility models using generalized additive models
.
Geomorphology
129
(
3–4
),
376
386
.
Guo
J.
,
Zhou
J.
,
Qin
H.
,
Zou
Q.
&
Li
Q.
2011
Monthly streamflow forecasting based on improved support vector machine model
.
Expert Systems with Applications
38
(
10
),
13073
13081
.
Halwatura
D.
&
Najim
M.
2013
Application of the HEC-HMS model for runoff simulation in a tropical catchment
.
Environmental Modelling & Software
46
,
155
162
.
Hirschberg
J.
&
Manning
C. D.
2015
Advances in natural language processing
.
Science
349
(
6245
),
261
266
.
Hochreiter
S.
&
Schmidhuber
J.
1997
Long short-term memory
.
Neural Computation
9
(
8
),
1735
1780
.
Jain
A.
&
Kumar
A. M.
2007
Hybrid neural network models for hydrologic time series forecasting
.
Applied Soft Computing
7
(
2
),
585
592
.
Jakeman
A.
&
Hornberger
G.
1993
How much complexity is warranted in a rainfall-runoff model?
Water Resources Research
29
(
8
),
2637
2649
.
Kalita
D.
,
Borbora
K. A.
&
Nath
D.
2022
Use of Bidirectional Long Short Term Memory in Spoken Word Detection with reference to the Assamese language
.
Indian Journal of Science and Technology
15
(
27
),
1364
1371
.
Kautz
T.
,
Groh
B. H.
,
Hannink
J.
,
Jensen
U.
,
Strubberg
H.
&
Eskofier
B. M.
2017
Activity recognition in beach volleyball using a deep convolutional neural network: Leveraging the potential of deep learning in sports
.
Data Mining and Knowledge Discovery
31
,
1678
1705
.
Kim
J.
,
Lee
J. K.
&
Lee
K. M.
2016
Deeply-recursive convolutional network for image super-resolution. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 27–30, 2016, Las Vegas, NV. IEEE, New York, pp. 1637–1645
.
Kimball
B.
1974
Smoothing data with Fourier transformations 1
.
Agronomy Journal
66
(
2
),
259
262
.
Krause
S.
,
Lewandowski
J.
,
Dahm
C. N.
&
Tockner
K.
2015
Frontiers in real-time ecohydrology – A paradigm shift in understanding complex environmental systems
.
Ecohydrology
8
(
4
),
529
537
.
Kumar
D.
,
Singh
A.
,
Samui
P.
&
Jha
R. K.
2019
Forecasting monthly precipitation using sequential modelling
.
Hydrological Sciences Journal
64
(
6
),
690
700
.
Legesse
D.
,
Vallet-Coulomb
C.
&
Gasse
F.
2003
Hydrological response of a catchment to climate and land use changes in tropical Africa: Case study South Central Ethiopia
.
Journal of Hydrology
275
(
1–2
),
67
85
.
Li
W.
,
Kiaghadi
A.
&
Dawson
C.
2021b
High temporal resolution rainfall–runoff modeling using long-short-term-memory (LSTM) networks
.
Neural Computing and Applications
33
,
1261
1278
.
Liu
M.
,
Huang
Y.
,
Li
Z.
,
Tong
B.
,
Liu
Z.
,
Sun
M.
,
Jiang
F.
&
Zhang
H.
2020
The applicability of LSTM-KNN model for real-time flood forecasting in different climate zones in China
.
Water
12
(
2
),
440
.
Mao
G.
,
Wang
M.
,
Liu
J.
,
Wang
Z.
,
Wang
K.
,
Meng
Y.
,
Zhong
R.
,
Wang
H.
&
Li
Y.
2021
Comprehensive comparison of artificial neural networks and long short-term memory networks for rainfall-runoff simulation
.
Physics and Chemistry of the Earth, Parts A/B/C
123
,
103026
.
Mehdizadeh
S.
,
Fathian
F.
&
Adamowski
J. F.
2019
Hybrid artificial intelligence-time series models for monthly streamflow modeling
.
Applied Soft Computing
80
,
873
887
.
Miotto
R.
,
Wang
F.
,
Wang
S.
,
Jiang
X.
&
Dudley
J. T.
2018
Deep learning for healthcare: Review, opportunities and challenges
.
Briefings in Bioinformatics
19
(
6
),
1236
1246
.
Moon
H.
,
Yoon
S.
&
Moon
Y.
2023
Urban flood forecasting using a hybrid modeling approach based on a deep learning technique
.
Journal of Hydroinformatics
25
(
2
),
593
610
.
Ongsulee
P.
2017
Artificial Intelligence, Machine Learning and Deep Learning
.
IEEE
,
New York
, pp.
1
6
.
Pan
X.
,
Liu
Z.
,
He
C.
&
Huang
Q.
2022
Modeling urban expansion by integrating a convolutional neural network and a recurrent neural network
.
International Journal of Applied Earth Observation and Geoinformation
112
,
102977
.
Piccialli
F.
,
Di Somma
V.
,
Giampaolo
F.
,
Cuomo
S.
&
Fortino
G.
2021
A survey on deep learning in medicine: Why, how and when?
Information Fusion
66
,
111
137
.
Ramírez
J. A.
2000
Prediction and modeling of flood hydrology and hydraulics
. In:
Inland flood hazards: Human, riparian and aquatic communities
(Wohl, E. E., ed.). Cambridge University Press, Cambridge
, p.
498
.
Reichstein
M.
,
Camps-Valls
G.
,
Stevens
B.
,
Jung
M.
,
Denzler
J.
&
Carvalhais
N.
2019
Deep learning and process understanding for data-driven Earth system science
.
Nature
566
(
7743
),
195
204
.
Ruelland
D.
,
Ardoin-Bardin
S.
,
Billen
G.
&
Servat
E.
2008
Sensitivity of a lumped and semi-distributed hydrological model to several methods of rainfall interpolation on a large basin in West Africa
.
Journal of Hydrology
361
(
1–2
),
96
117
.
Santikari
V. P.
&
Murdoch
L. C.
2018
Including effects of watershed heterogeneity in the curve number method using variable initial abstraction
. Hydrology and Earth System Sciences
22
(
9
),
4725
4743
.
Scharffenberg
W.
,
Ely
P.
,
Daly
S.
,
Fleming
M.
&
Pak
J.
2010
Hydrologic Modeling System (HEC-HMS): Physically-Based Simulation Components. In: 2nd joint federal interagency conference, Las Vegas, NV. USGS, Reston, VA.
Shen
C.
,
Laloy
E.
,
Elshorbagy
A.
,
Albert
A.
,
Bales
J.
,
Chang
F.-J.
,
Ganguly
S.
,
Hsu
K.-L.
,
Kifer
D.
&
Fang
Z.
2018
HESS opinions: Incubating deep-learning-powered hydrologic science advances as a community
.
Hydrology and Earth System Sciences
22
(
11
),
5639
5656
.
Skøien
J. O.
,
Blöschl
G.
&
Western
A. W.
2003
Characteristic space scales and timescales in hydrology
.
Water Resources Research
39
(
10
),
11
30
.
Suryoputro
N.
,
Suhardjono
,
Soetopo
W.
&
Suhartanto
E.
2017
Calibration of infiltration parameters on hydrological tank model using runoff coefficient of rational method
.
AIP Conference Proceedings
1887
,
020056
.
Tahmasebi
P.
,
Kamrava
S.
,
Bai
T.
&
Sahimi
M.
2020
Machine learning in geo-and environmental sciences: From small to large scale
.
Advances in Water Resources
142
,
103619
.
Tan
K. L.
,
Lee
C. P.
,
Anbananthen
K. S. M.
&
Lim
K. M.
2022
RoBERTa-LSTM: a hybrid model for sentiment analysis with transformer and recurrent neural network
.
IEEE Access
10
,
21517
21525
.
Tang
Y.
2016
TF. Learn: TensorFlow's high-level module for distributed machine learning. arXiv preprint arXiv:1612.04251
.
Thornton
P. E.
,
Running
S. W.
&
White
M. A.
1997
Generating surfaces of daily meteorological variables over large regions of complex terrain
.
Journal of Hydrology
190
(
3–4
),
214
251
.
Torfi
A.
,
Shirvani
R. A.
,
Keneshloo
Y.
,
Tavaf
N.
&
Fox
E. A.
2020
Natural language processing advancements by deep learning: A survey. arXiv preprint arXiv:2003.01200
.
Wang
Y.
,
Wu
H.
,
Zhang
J.
,
Gao
Z.
,
Wang
J.
,
Philip
S. Y.
&
Long
M.
2022
PredRNN: A recurrent neural network for spatiotemporal predictive learning
.
IEEE Transactions on Pattern Analysis and Machine Intelligence
45
(
2
),
2208
2225
.
Xiang
Z.
,
Yan
J.
&
Demir
I.
2020
A rainfall-runoff model with LSTM-based sequence-to-sequence learning
.
Water Resources Research
56
(
1
),
e2019WR025326
.
Xie
D.
,
Zhang
L.
&
Bai
L.
2017
Deep learning in visual computing and signal processing
.
Applied Computational Intelligence and Soft Computing
2017
,
1320780
.
Yan
L.
,
Lei
Q.
,
Jiang
C.
,
Yan
P.
,
Ren
Z.
,
Liu
B.
&
Liu
Z.
2022
 Climate-informed monthly runoff prediction model using machine learning and feature importance analysis
.
Frontiers in Environmental Science
10
,
1049840
.
Yuan
X.
,
Zhou
J.
,
Huang
B.
,
Wang
Y.
,
Yang
C.
&
Gui
W.
2019
Hierarchical quality-relevant feature representation for soft sensor modeling: A novel deep learning strategy
.
IEEE Transactions on Industrial Informatics
16
(
6
),
3721
3730
.
This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (CC BY 4.0), which permits copying, adaptation and redistribution, provided the original work is properly cited (http://creativecommons.org/licenses/by/4.0/).