In the context of global climate change and the continuous development of urban areas, rainfall-inundation modeling is a common approach that provides critical support for the protection and early warning of urban waterlogging protection. The present study conducts a data-driven model for hourly urban rainfall-inundation depth prediction, which is based on a gated recurrent unit (GRU) neural network and uses the simulated annealing (SA) algorithm for the hyperparameter optimization of GRU, namely the SA-GRU model. To verify the performance of the proposed model, backpropagation, long short-term memory (LSTM), and bidirectional LSTM (BiLSTM) neural networks are set as benchmarks. Results show that the SA-GRU has high accuracy in the case of short-term inundation prediction, with the Nash–Sutcliffe efficiency from 0.999 to 0.596 for the 1-h-ahead to 8-h-ahead predictions. And further research reveals that the SA-GRU integrates the significant optimization of SA, with an average 20% reduction of the root mean square error within the first eight prediction periods, and the efficient training speed of GRU, with 23.7% faster than LSTM and 44.2% faster than BiLSTM. In conclusion, the SA-GRU excels in urban inundation prediction, demonstrating its value in flood management and decision-making.

  • A GRU-based neural network was used for urban waterlogging modeling to make inundation depth predictions for 1–12 h.

  • A simulated annealing (SA) algorithm was used as the hyperparameter optimization method to boost the performance of GRU.

  • The SA-GRU model has better performance rather than SA-BP, SA-LSTM, and SA-BiLSTM within its prediction period threshold (8 h).

Over the past few decades, increasing urban flooding events have caused more severe disasters with the explosion of urbanization and the influence of climate change around the world (De Silva & Kawasaki 2020; Kao et al. 2021; Valeh et al. 2021; Wang J. et al. 2021; Wang P. et al. 2021). Urban flooding usually results in the loss of economic assets and livelihoods, disrupts the human daily routine and urban business activities, and seriously plagues urban development (Galloway et al. 2018; Wang et al. 2020). As the fastest-growing urban agglomerations in the world, many Asian cities experience the rapid transformation of the underlying surface, leading to a significant decrease in the capacity of water seepage and greater urban flooding risks in most areas (Yi & Yang 2014; Luo et al. 2018; Kao et al. 2021). For instance, China has experienced a rising incidence of urban flooding since 2010 (Govt 2020).

To cope with urban flooding and reduce losses, urban rainfall-flood models are used for inundation simulation and prediction. These models can be categorized as physical-based and data-driven models (Yuan et al. 2018; Tamiru & Dinka 2021). The biggest difference between the two types of models is that physical-based models are composed of equations based on physical phenomena, while data-driven models depend on data and statistical methods (Li et al. 2021). One modeling strategy for flood simulation and prediction in urban catchments involves coupling meteorological and urban hydrological models (Nguyen & Bae 2020), which facilitates a comprehensive simulation of the entire process from rainfall initiation to urban flood evolution. Numerous physically based models, such as the Hydrologic Engineering Center's River Analysis System (HEC-RAS), have been developed using this methodology and are widely applied in urban flood inundation simulation and risk analysis (Rangari et al. 2019; Feng et al. 2021; Madhuri et al. 2021).

Despite the continued dominance of traditional physically based models, they have encountered several critical challenges that have hindered their progress over time. One key limitation is the high computational demands associated with the complex processes and phenomena they represent. This can create difficulties for large-scale simulations or real-time flood forecasting, potentially requiring significant computational resources and restricting their applicability in certain situations (Kao et al. 2021). Moreover, physically based models typically necessitate extensive input data for calibration and validation, such as topographic, land use, and rainfall data. Acquiring high-quality, consistent data can be challenging, particularly in developing countries or remote areas, which can impact the accuracy and effectiveness of physical models in these contexts. Furthermore, physically based models involve numerous parameters that require calibration for the accurate representation of site-specific conditions (Herrera et al. 2022). The calibration process can be time-consuming and computationally demanding, and the models can be sensitive to parameter values, potentially introducing uncertainties in the results if calibration is not performed accurately.

With the increasing approaches to data acquisition and the growing information technology, data-driven models thrive rapidly and are gradually applied to the field of hydrology. Since the 1990s, artificial neural networks (ANNs) have been used for hydrological modeling (Hsu et al. 1995; Dawson & Wilby 1998; Abrahart et al. 2004; Asadi et al. 2013; Kan et al. 2015; Chang et al. 2018; Kao et al. 2021). To deal with time series tasks better, a class of ANNs was designed, named recurrent neural networks (RNNs, Rumelhart et al. 1986; Li et al. 2021). Depending on the unit structure, RNNs are divided into different types, including long short-term memory (LSTM) and its variant gated recurrent unit (GRU). To date, LSTM has obtained significant achievements in various directions in the field of hydrological forecasting, such as precipitation forecasting (Kumar et al. 2019), runoff and water-level forecasting (Kratzert et al. 2018; Widiasari et al. 2018; Yuan et al. 2018; Kao et al. 2020; Xiang et al. 2020; Zuo et al. 2020; Li et al. 2021), flood disaster early warning (de la Fuente et al. 2019), and groundwater forecasting (Zhang et al. 2018). However, the numerous trainable parameters of LSTM can lead to prolonged model training times (Wang J. et al. 2022). As a result, a simplified version of the LSTM architecture, the GRU, was introduced (Cho et al. 2014). The GRU has only two gates (update and reset) and lacks a separate memory cell, resulting in fewer parameters, faster training times, and lower computational requirements compared to LSTMs (Fu et al. 2016; Chen et al. 2022). Therefore, the potential application and value of GRU attract the attention of some hydrologists and researchers recently. Gao et al. (2020) compared LSTM and the GRU in short-term runoff predictions and concluded that the GRU might be the preferred method. In terms of the direction of groundwater forecasting and water quality prediction, the latest research also shows good application prospects for the GRU (Gharehbaghi et al. 2022; Mei et al. 2022).

Apart from the unit structure, the performance of a neural network largely depends on its hyperparameters that define the model architecture (Kuhn & Johnson 2013; Roy et al. 2014), such as the number of layers and the number of neurons in each layer. Therefore, tuning hyperparameters is essential for building an effective neural network model (Gressling 2020). Traditional empirical hyperparameter optimization involves a manual trial-and-error process, where researchers experiment with different configurations of hyperparameters based on their experience, intuition, or previous research findings. This approach can be time-consuming and may not guarantee optimal results, as it relies on human judgment and expertise. In contrast, systematic, algorithm-driven search processes employ algorithms that systematically explore the hyperparameter space to find the best configuration of hyperparameters, independent of human intervention. These algorithm-driven methods have the advantage of being more objective, efficient, and reduce human effort while enhancing model performance and reproducibility (Yang & Shami 2020). Simulated annealing (SA) is a modern heuristic that has been extensively applied to solve optimization problems (Avello et al. 2005). Wang Z. et al. (2022) applied SA with a particle swarm algorithm to obtain the optimum allocation schemes of water resources at precipitation frequencies. Wang J. et al. (2021) used the driving force-pressure-state-impact-response-management framework model and the SA-projection pursuit model to calculate the water resource carrying capacity to determine the main factors affecting the lake ecosystem. Hosseini et al. (2020) used SA to eliminate redundant variables for the temporal flash-flood forecasting models building to enhance the accuracy of assessing hazardous areas. However, compared with other optimization algorithms (Yuan et al. 2018; Yang & Shami 2020; Xu et al. 2022), there are few pieces of research using SA for hyperparameter optimization in neural networks.

The purpose of this study is to develop the GRU neural network boosted by an SA-based hyperparameter optimization algorithm (SA-GRU) for urban flood inundation depth prediction. Benefiting from the advantages of the accuracy and speed of GRU in time series forecasting, especially with the improvement of SA-based hyperparameter optimization, the proposed model has great potential to break through the technical bottleneck that traditional physical-based models have. Meanwhile, the feasibility verification done in this study can also be preliminary work to make modern algorithms a solution worth considering for real-time urban flood-inundation prediction. The dataset employed for neural networks comprises rainfall time series and inundation depth time series. Most of the rainfall data was sourced from the local administration office in Hefei City, China, while a small portion was based on designed rainfall. Inundation data was generated by the HEC-RAS model, which has been validated as a feasible method for simulating urban flood events in real projects. To verify the performance of the proposed model in various prediction periods, we tested the SA-GRU for the next 1–12 h of inundation depth prediction separately. Furthermore, several other neural networks were set as benchmarks. On the one hand, the original GRU with the empirical hyperparameter configuration was compared with the SA-GRU to explore the improvement of SA on GRU. On the other hand, to further validate the superiority of SA-GRU for urban flood-inundation depth prediction, SA was also adopted to optimize the hyperparameters of backpropagation (BP) neural network, LSTM, and bidirectional LSTM (BiLSTM), respectively, and their performance was compared in each prediction period.

The rest of this study is organized as follows. Section 2 overviews the study area and dataset. Section 3 designs the framework of experiments and introduces the relevant methods. Section 4 presents the results and the discussions on the performance of the proposed SA-GRU model for urban inundation depth prediction. Eventually, conclusions are then drawn in Section 5.

Study area and dataset

Hefei is the capital and the biggest city of Anhui province in East China and is located at 31.87°N, 117.28°E. The total area of Hefei is 11,445 km2. The region of Hefei features a humid subtropical climate with four distinct seasons. Its annual precipitation varies considerably and is slightly over 1,000 mm and mostly concentrated in summer. In history, Hefei is a city that suffers from frequent floods. Since 1949, there have been 18 floods in Hefei, of which the city suffered 4 major floods. As shown in Figure 1, the study area was selected in the central urban area of Hefei.
Figure 1

Base map. (a) Study area, (b) Digital Elevation Model (DEM), and (c) land cover.

Figure 1

Base map. (a) Study area, (b) Digital Elevation Model (DEM), and (c) land cover.

Close modal

To predict future inundation depths based on historical rainfall-inundation data without involving other features, the dataset in this study consists of rainfall time series and inundation depth time series. The rainfall time series were composed of observed and designed rainfall events, and due to the lack of observed inundation data, the inundation time series were generated by the HEC-RAS model based on the rainfall time series and relevant topographic information. Among these, the observed rainfall data were acquired from http://112.30.184.124:18001/ExternalAdmin, spanning from 2010 to 2021. As the relationship between peak rainfall intensity (PRI) and maximum flood depth exhibits an exponential trend (Sahoo & Sreeja 2016), rainfall events were selected with a PRI greater than 10 mm/h to yield more pronounced inundation processes. Ultimately, the rainfall time series included 57 observed rainfall events and 8 designed rainfall events. After applying several data preprocessing methods (detailed in Section 3.4.3.), the resulting dataset contained 65 sets of rainfall-inundation time series at 15-min intervals. The entire rainfall time series includes a total of 10,765-time instants, with the longest-lasting rainfall event consisting of 376-time instants (nearly 4 days). Since it becomes challenging to forecast inundation with excessively long prediction periods, the prediction period threshold was tentatively set to 12 h, resulting in every inundation time series being 12 h longer than the rainfall time series. Consequently, the dataset contains a total of 13,885 inundation time instants (10,765 + 12 × 4 × 65). A comprehensive overview of the dataset is provided in Table 1.

Table 1

Dataset overview

RainfallInundation depth
Total events 65 65 
Data source Observed (57) and designed (8) Generated 
Unit Millimeter per hour (mm/h) Meter (m) 
Data amount 10,765 time instants 13,885 time instants 
Longest single event 376 time instants (94 h) 424 time instants (106 h) 
Average value 3.4 mm/h 0.42 m 
Minimum value 0 mm/h 0 m 
Maximum value 62.5 mm/h 2.14 m 
Standard deviation 6.6 mm/h 0.36 m 
RainfallInundation depth
Total events 65 65 
Data source Observed (57) and designed (8) Generated 
Unit Millimeter per hour (mm/h) Meter (m) 
Data amount 10,765 time instants 13,885 time instants 
Longest single event 376 time instants (94 h) 424 time instants (106 h) 
Average value 3.4 mm/h 0.42 m 
Minimum value 0 mm/h 0 m 
Maximum value 62.5 mm/h 2.14 m 
Standard deviation 6.6 mm/h 0.36 m 

Experimental framework

Within the proposed SA-GRU model, the GRU-based neural network is the functional core for inundation depth prediction and the SA algorithm is the hypermodel utilized to search for the optimal hyperparameter configuration of the GRU. As shown in Figure 2, three experiments with different focuses are designed to evaluate the proposed SA-GRU model:
  • 1.

    Evaluation of the SA-GRU across various prediction periods: To verify the feasibility of SA-GRU in urban inundation depth prediction and determine the predictable duration, this experiment presented the forecasting results of inundation depth based on the SA-GRU in different prediction periods. Through several tests, 1–12 h was found as an appropriate range of prediction periods.

  • 2.

    Comparison between the original GRU and the SA-GRU: To quantify and analyze the improvement of SA on the predictive capacity of the GRU, an original GRU was developed as the benchmark with the empirical hyperparameter configuration given by Gao et al. (2020). The two models were compared in all 12 prediction periods. In addition, since the hyperparameter optimization progress usually requires massive iterations in which massive corresponding GRU networks with different hyperparameter configurations were developed, another method was added to verify the effect of SA on the GRU by analyzing the differences between the optimal and the average performance of these GRU networks.

  • 3.

    Exploration of the effect of SA on different neural networks: In this experiment, the BP neural network and other two RNNs (LSTM and BiLSTM) were all developed with the SA-based hyperparameter optimization algorithm, building SA-BP, SA-LSTM, and SA-BiLSTM, respectively. Based on the type of neural networks, two aspects were researched with these benchmarks. On the one hand, SA-BP was compared with the SA-GRU to verify whether the GRU still takes the advantage of RNNs in the specific urban inundation depth prediction as long as other time series prediction tasks. On the other hand, to explore the differences between the mentioned three RNNs, SA-LSTM and SA-BiLSTM were compared to the SA-GRU.

Figure 2

Overview of the comprehensive experimental framework. Based on the hyperparameter optimization strategy, benchmarks include an empirically optimized model (the original GRU) and three SA-optimized models (SA-LSTM, SA-BiLSTM, and SA-BP).

Figure 2

Overview of the comprehensive experimental framework. Based on the hyperparameter optimization strategy, benchmarks include an empirically optimized model (the original GRU) and three SA-optimized models (SA-LSTM, SA-BiLSTM, and SA-BP).

Close modal

LSTM and GRU neural networks

LSTM neural network

LSTM is a recurrent neural network with memory structures for learning both short-term and long-term information (Xiang et al. 2020), which has strong nonlinear fitting capabilities and effectively solves the vanishing and exploding gradient problems of standard RNN (Wang j. et al. 2022). As shown in Figure 3, standard LSTM has a cell that can save information over arbitrary time intervals and three gates (the forget, the input, and the output gates) that direct the information flow into and out of the cell (Kao et al. 2021). The forget gate controls how much of the previous state is to be forgotten. The input gate controls how much of the new state can be added to the current state. The output gate controls how much of the current state is going to transfer to the next state.
Figure 3

Architecture of a single LSTM unit.

Figure 3

Architecture of a single LSTM unit.

Close modal
At time step t, six model variables are updated in a single LSTM unit. The calculation formula for each variable is shown as follows.
(1)
(2)
(3)
(4)
(5)
(6)
where ft is the activation vector of the forget gate; it is the activation vector of the input gate; ot is the activation vector of the output gate; gt is the activation vector of the cell input; xt is the input vector to the LSTM unit; W and b are the weight matrices and bias vector that are trainable variables; and tanh are the sigmoid function and the hyperbolic tangent function, generating values between (0,1) and [−1,1], respectively; Ct and Ct−1 are the current and the previous state vector; ht is the hidden state vector.

Besides standard LSTM, BiLSTM is also widely used, which consists of two LSTMs, one taking the input in a forward direction and the other in a backward direction. For some specific tasks, BiLSTM effectively increases the amount of information and learns deeper than one-directional LSTM. Therefore, BiLSTM is considered another benchmark of GRU in this study.

GRU neural network

As shown in Figure 4, the main difference between the GRU and LSTM is that the GRU has only two gates: the reset gate and the update gate. Within the GRU, the reset gate controls how much of the previous state can be retained and the update gate determines how much of the previous state needs to pass along to the next state.
Figure 4

Architecture of a single GRU unit.

Figure 4

Architecture of a single GRU unit.

Close modal
As shown in Equations (7)–(10), at time step t, a single GRU unit has four model variables to compute less than six of LSTMs.
(7)
(8)
(9)
(10)
where is the activation vector of the reset gate and is the activation vector of the update gate. Other parameters are the same as in LSTM.

Compared with LSTM structurally, the GRU does not have an output gate and combines the input and the forget gates into a single update gate. Therefore, the GRU is less complex and executes faster than LSTM, whereas LSTM is more accurate on a larger sequence theoretically. In this study, firstly, the amount of dataset used for experiments is not that much. What's more, since the hyperparameters of neural networks were designed to be optimized by SA, a great number of iterations result in longer computation time and thus require faster training speed. So the GRU is considered the more appropriate neural network for the specific task of this study, especially if applied to real-time urban inundation depth prediction in the future.

SA algorithm

SA is a stochastic optimization algorithm inspired by the annealing process in metallurgy, where metals are heated to high temperatures and then slowly cooled to reduce defects and improve their structural properties (Kirkpatrick et al. 1983). The algorithm is particularly useful for finding approximate solutions to complex optimization problems, especially those with a large search space or numerous local optima (Jansen 2011). The key principles of SA can be summarized as follows.

  • 1.

    Initialization: Define an initial solution, an objective function (energy function) to evaluate the quality of solutions, a temperature schedule, and an initial temperature value. The temperature represents a control parameter that determines the degree of randomness allowed in the search process.

  • 2.

    Perturbation: Generate a new solution by applying a small, random modification to the current solution. This step is analogous to the movement of atoms in a heated metal.

  • 3.

    Evaluation: Calculate the change in the objective function between the current solution and the new solution.

  • 4.

    Acceptance: Accept or reject the new solution based on the change in the objective function and the current temperature. If the new solution is better, it is always accepted. If it is worse, it may still be accepted with a probability that decreases as the temperature decreases.

  • 5.

    Cooling: Decrease the temperature according to the temperature schedule. This gradual decrease in temperature is analogous to the slow cooling process in metallurgy, which allows the system to settle into a low-energy state.

  • 6.

    Iteration: Repeat the perturbation, evaluation, and acceptance steps until a stopping criterion is met, such as reaching a minimum temperature or a maximum number of iterations.

This acceptance and retention procedure allows the algorithm to explore the search space more effectively, as it helps to escape local optima and eventually converges to a global optimum or near-optimal solution. The gradual reduction in temperature ensures that the algorithm becomes more selective over time, accepting only better solutions as it approaches convergence.

In SA, the acceptance probability is given by the Metropolis criterion (Metropolis et al. 1953). It determines the likelihood of accepting a worse solution (S′) compared to the current solution (S). The acceptance probability, P, is calculated as follows:
(11)
(12)
where E represents the objective function or energy function, K is the Boltzmann constant, and T is the temperature. The exponential function ensures that the probability of accepting a worse solution decreases as the temperature T decreases or as the difference ΔE increases.
The flow of SA-based hyperparameter optimization is shown in Figure 5. The search space was defined as the value sets of key hyperparameters. The objective function E was set to the negative of Nash–Sutcliffe efficiency (NSE) coefficient (mentioned in Section 2.5.2) obtained by the neural network model on the validation set, as shown in Equation (13). Therefore, the objective function E ranges from [−1, +inf) and as the SA process progresses, a lower E value is expected.
(13)
Figure 5

Flowchart of SA algorithm for the hyperparameter optimization of neural networks.

Figure 5

Flowchart of SA algorithm for the hyperparameter optimization of neural networks.

Close modal

Next, the control parameters of SA are initialized. In this study, the initial temperature and the termination condition were set to 100° and 100 iterations, respectively. According to the Metropolis criterion, SA has a certain probability of escaping local minima, and the probability decreases as the temperature decreases, with the optimal solution gradually converging. Finally, the optimal hyperparameter configuration is obtained when the termination condition is reached.

A hybrid SA-GRU model for urban inundation depth prediction

Based on methods of GRU and the SA-based hyperparameter optimization, there are two major objects that request the definition for building the hybrid SA-GRU model: key hyperparameters (including their value sets) and evaluation criteria. In addition, several data preprocessing methods are commonly conducted before neural network calibration.

Key hyperparameters

According to the features of neural networks and the purpose of this study, three sets of key hyperparameters and their value sets are defined, as shown in Table 2. The first set is related to the architecture of GRU neural networks, including the number of hidden layers and the number of neurons in each layer. The second set comprises the minimum batch size and the number of epochs, representing the number of processed samples before updating the model and the number of complete passes through the entire calibration set, respectively (Soon et al. 2018). The minimum batch size depends on the resource requirements of the training process and the number of iterations. The number of epochs affects training results and time. If epochs are insufficient, the model is prone to underfit. Instead, if epochs are too much, it will lead to unnecessary extra execution time, even overfitting. To reduce optimization time, the maximum value of epochs is set to 100, which has been tested to be sufficient for convergence. Moreover, a suitable number of neurons in the input layer is also expected, since if the input sequence is too short or too long, that results in poor performance of models (Gao et al. 2020). Therefore, the length of the sequence, or the neurons of the input layer, is considered as the third set of hyperparameters, ranging from 4 to 12 h (corresponding neurons quadruple based on the 15-min intervals).

Table 2

Specifications of the hyperparameter to be optimized

HyperparametersValue sets
The number of hidden layers {1, 2, 3} 
The number of neurons in each layer {4, 8, 16, 32, 64, 128, 256} 
Mini-batch size {32, 64, 128, 256} 
The number of epochs {30, 40, 50, 60, 70, 80, 90, 100} 
The length of the input sequence {16, 20, 24, 28, 32, 36, 40, 44, 48} 
HyperparametersValue sets
The number of hidden layers {1, 2, 3} 
The number of neurons in each layer {4, 8, 16, 32, 64, 128, 256} 
Mini-batch size {32, 64, 128, 256} 
The number of epochs {30, 40, 50, 60, 70, 80, 90, 100} 
The length of the input sequence {16, 20, 24, 28, 32, 36, 40, 44, 48} 

Evaluation criteria

To evaluate the performance of the proposed SA-GRU model and benchmarks, the NSE coefficient, the root mean square error (RMSE), and the mean absolute error (MAE) are adopted as evaluation criteria. Related equations are as follows:
(14)
(15)
(16)
where is the predicted inundation depth value at the time step i; is the objective inundation depth value at the time step i; is the average value of the objective inundation depths and n is the data amount of the sequence.

The NSE is a dimensionless efficiency index and was first proposed by Nash and Sutcliffe in 1970 to evaluate hydrological simulations. The value of NSE ranges from −∞ to 1, the closer to 1, the better the models. Research shows that the NSE has a higher sensitivity to peak values (Waseem et al. 2008), which mainly helps to evaluate the accuracy of models for the maximum inundation depth prediction.

RMSE and MAE are dimensional (in meters in this study) and mathematical criteria that evaluate distances between the objective and the predicted values. Both errors are greater than 0. It is suggested that MAE is less sensitive to outliers and that it is an unambiguous measure of average error magnitude compared with RMSE (Willmott & Matsuura 2005).

In most hydrological model evaluations, model performance can be categorized into four levels based on NSE values: (0.75, 1], (0.65, 0.75], (0.5, 0.65], and (−∞, 0.5]. To assess model performance more precisely, a six-level performance rating system, which is applicable to this study and based on NSE, RMSE, and MAE, is employed, as shown in Table 3.

Table 3

Refined model evaluation system based on NSE, RMSE, and MAE

NSE rangeRMSE rangeMAE rangePerformance rating
(0.9, 1] (0, 0.10] (0, 0.08] Excellent 
(0.75, 0.9] (0.10, 0.20] (0.08, 0.15] Very good 
(0.65, 0.75] (0.20, 0.25] (0.15, 0.20] Good 
(0.6, 0.65] (0.25, 0.30] (0.20, 0.25] Satisfactory 
(0.5, 0.6] (0.30, 0.35] (0.25, 0.30] Acceptable 
(−∞, 0.5] (0.35, +∞] (0.30, +∞] Unacceptable 
NSE rangeRMSE rangeMAE rangePerformance rating
(0.9, 1] (0, 0.10] (0, 0.08] Excellent 
(0.75, 0.9] (0.10, 0.20] (0.08, 0.15] Very good 
(0.65, 0.75] (0.20, 0.25] (0.15, 0.20] Good 
(0.6, 0.65] (0.25, 0.30] (0.20, 0.25] Satisfactory 
(0.5, 0.6] (0.30, 0.35] (0.25, 0.30] Acceptable 
(−∞, 0.5] (0.35, +∞] (0.30, +∞] Unacceptable 

Model development

Figure 6 illustrates the steps of SA-GRU development and the flowchart of data processing. The methods that generate the rainfall-inundation dataset from raw materials are presented in Section 2.2., including data filtering and inundation depth data generation. For data augmentation, the dataset was interpolated from original 1-h intervals to 15-min intervals and thus quadrupled. Linear interpolation was selected as the interpolation method to avoid the occurrence of data with negative values. Moreover, due to the order of magnitude gaps in rainfall and inundation depth values, it was necessary to normalize the dataset to reduce the impact of magnitude and improve the speed of model convergence at the same time. The normalization method based on min–max scaling is shown as follows.
(17)
where is the normalized value and is the raw input.
Figure 6

Flowchart of the overall SA-GRU model development procedure.

Figure 6

Flowchart of the overall SA-GRU model development procedure.

Close modal

As indicated in Table 4, the dataset was split into calibration, validation, and test datasets. Utilizing the model development methods, experimental framework, and evaluation criteria described earlier, the SA-GRU model and benchmark models were developed and evaluated.

Table 4

Dataset split proportions

DatasetsData ratio (%)
Calibration set 72 
Validation set 18 
Test set 10 
DatasetsData ratio (%)
Calibration set 72 
Validation set 18 
Test set 10 

All experiments are conducted with the hardware equipment Intel(R) Core (TM) i5-1135G7 CPU @2.40 GHz, RAM 16.00 GB, and NVIDIA GeForce MX450. The programming language for data processing and model development is Python 3.9.7, with related packages including Numpy 1.19.5, Pandas 1.4.1, Matplotlib 3.5.1, TensorFlow 2.6.0, and Hyperopt 0.2.7.

Results of the SA-GRU in different prediction periods

Table 5 presents the optimal hyperparameter configuration in each prediction period. The result shows that the optimal length of the input sequence is 20 neurons (5 h), which verifies that it is the appropriate one, not the longer input sequence length that can be helpful to improve the performance of GRU. In terms of the number of hidden layers, it is found that networks in 10 of 12 prediction periods are optimized to the architecture of one single hidden layer with 64 neurons. A one hidden layer network is likely to fit better than multilayer networks for rainfall-inundation prediction. As for the number of epochs, it is shown that 80–100 can be a suitable range. The optimal minimum batch size is 32.

Table 5

The results of the optimal hyperparameter configuration

Prediction period (h)The length of the input sequenceThe number of hidden layersThe number of neurons in each layerThe number of epochsMinimum batch size
20 (5 h) [64] 90 32 
20 (5 h) [64, 32] 90 32 
16 (4 h) [64, 32] 90 32 
24 (6 h) [64] 100 32 
16 (4 h) [64] 90 32 
24 (6 h) [32] 80 32 
20 (5 h) [64] 90 32 
16 (4 h) [64] 80 32 
32 (8 h) [64] 100 32 
10 20 (5 h) [32] 90 32 
11 20 (5 h) [64] 90 32 
12 24 (6 h) [64] 80 32 
Prediction period (h)The length of the input sequenceThe number of hidden layersThe number of neurons in each layerThe number of epochsMinimum batch size
20 (5 h) [64] 90 32 
20 (5 h) [64, 32] 90 32 
16 (4 h) [64, 32] 90 32 
24 (6 h) [64] 100 32 
16 (4 h) [64] 90 32 
24 (6 h) [32] 80 32 
20 (5 h) [64] 90 32 
16 (4 h) [64] 80 32 
32 (8 h) [64] 100 32 
10 20 (5 h) [32] 90 32 
11 20 (5 h) [64] 90 32 
12 24 (6 h) [64] 80 32 

Table 6 demonstrates the performance of the SA-GRU model in predicting urban inundation depth across 1–12-h prediction periods. The model is assessed using both validation and test sets based on NSE, RMSE, MAE, and performance rating.

Table 6

The evaluation of SA-GRU in prediction periods from 1 to 12 h

Prediction period (h)DatasetNSERMSE(m)MAE (m)Performance rating
Validation set 0.999 0.03 0.01 Excellent 
Test set 0.999 0.03 0.01 Excellent 
Validation set 0.995 0.03 0.02 Excellent 
Test set 0.995 0.03 0.02 Excellent 
Validation set 0.978 0.06 0.04 Excellent 
Test set 0.976 0.06 0.04 Excellent 
Validation set 0.943 0.10 0.07 Excellent 
Test set 0.940 0.10 0.07 Excellent 
Validation set 0.853 0.16 0.10 Very good 
Test set 0.845 0.16 0.10 Very good 
Validation set 0.768 0.20 0.13 Very good 
Test set 0.751 0.20 0.13 Very good 
Validation set 0.657 0.24 0.17 Good 
Test set 0.643 0.24 0.17 Satisfactory 
Validation set 0.596 0.26 0.17 Acceptable 
Test set 0.582 0.26 0.17 Acceptable 
Validation set 0.493 0.29 0.20 Unacceptable 
Test set 0.489 0.29 0.19 Unacceptable 
10 Validation set 0.418 0.30 0.21 Unacceptable 
Test set 0.399 0.31 0.21 Unacceptable 
11 Validation set 0.371 0.30 0.21 Unacceptable 
Test set 0.335 0.31 0.22 Unacceptable 
12 Validation set 0.212 0.33 0.23 Unacceptable 
Test set 0.197 0.34 0.23 Unacceptable 
Prediction period (h)DatasetNSERMSE(m)MAE (m)Performance rating
Validation set 0.999 0.03 0.01 Excellent 
Test set 0.999 0.03 0.01 Excellent 
Validation set 0.995 0.03 0.02 Excellent 
Test set 0.995 0.03 0.02 Excellent 
Validation set 0.978 0.06 0.04 Excellent 
Test set 0.976 0.06 0.04 Excellent 
Validation set 0.943 0.10 0.07 Excellent 
Test set 0.940 0.10 0.07 Excellent 
Validation set 0.853 0.16 0.10 Very good 
Test set 0.845 0.16 0.10 Very good 
Validation set 0.768 0.20 0.13 Very good 
Test set 0.751 0.20 0.13 Very good 
Validation set 0.657 0.24 0.17 Good 
Test set 0.643 0.24 0.17 Satisfactory 
Validation set 0.596 0.26 0.17 Acceptable 
Test set 0.582 0.26 0.17 Acceptable 
Validation set 0.493 0.29 0.20 Unacceptable 
Test set 0.489 0.29 0.19 Unacceptable 
10 Validation set 0.418 0.30 0.21 Unacceptable 
Test set 0.399 0.31 0.21 Unacceptable 
11 Validation set 0.371 0.30 0.21 Unacceptable 
Test set 0.335 0.31 0.22 Unacceptable 
12 Validation set 0.212 0.33 0.23 Unacceptable 
Test set 0.197 0.34 0.23 Unacceptable 

In the initial prediction periods of 1–4 h, the SA-GRU model performs exceptionally well for both validation and test sets. The NSE values are 0.999 for both sets, with low RMSE (0.03) and MAE (0.01) values. As the prediction period extends to 5 and 6 h, the model maintains a very good performance level on both sets, with a slight decrease in the test set performance. At the 6-h prediction period, the NSE value for the validation set is 0.768, while it is 0.751 for the test set.

For prediction periods of 7 and 8 h, the performance of SA-GRU shows good and acceptable levels, respectively, on both sets. The NSE values experience a significant reduction, and the RMSE and MAE values increase. At the 8-h prediction period, the NSE value for the validation set is 0.596, while it is 0.582 for the test set. When the prediction period surpasses 8 h (9–12 h), the performance of SA-GRU is considered unacceptable on both sets. For instance, at the 9-h and 12-h prediction periods, the NSE values for the test set are 0.489 and 0.197, respectively.

As illustrated in Figure 7, a comparison between the predicted inundation process by SA-GRU and the target inundation process by HEC-RAS for specific rainfall events is provided. Figure 7(a) represents an observed rainfall event, while Figure 7(b) represents a designed rainfall event. Results show that the model exhibits excellent short-term prediction capabilities for both events. However, as the prediction period extends to 8 and 12 h, the accuracy of SA-GRU declines, and a systematic error becomes more prominent. This error is likely due to the increasing difficulty in capturing future rainfall information with longer prediction periods. The trends observed in both observed and designed rainfall events also demonstrate that the SA-GRU model performs well in short-term predictions but faces limitations in longer prediction periods.
Figure 7

Comparison of inundation hydrographs for prediction periods of 1, 4, 8, and 12h. (a) Observed rainfall events and (b) designed rainfall events.

Figure 7

Comparison of inundation hydrographs for prediction periods of 1, 4, 8, and 12h. (a) Observed rainfall events and (b) designed rainfall events.

Close modal

In conclusion, the SA-GRU model excels in providing highly accurate predictions for urban inundation depth in the short term, showcasing its strength in these crucial timeframes. Despite the decreasing precision with extended prediction periods, the SA-GRU model continues to offer effective solutions for short-term rainfall-inundation predictions.

Comparison between the original GRU and the SA-GRU

The experiment aims to evaluate the improvement of SA on the GRU. The original GRU is set as the benchmark, with the empirical hyperparameter configuration given by Gao et al. (2020).

Table 7 shows the NSEs, RMSEs, and MAEs of the original GRU in prediction periods from 1 to 12 h. Overall, the proposed SA-GRU model performs better than the original GRU model in 11 of 12 prediction periods, except for the 1-h-ahead prediction that both models get NSEs as high as 0.999. As shown in Figure 8, the numbers in red represent the differences between the SA-GRU and the original GRU. The optimization effect of SA becomes more pronounced with the extension of the prediction period. As the prediction period increases to 4 h, the improvements in NSE, RMSE, and MAE become more significant, with NSE increasing by 0.08, and RMSE and MAE decreasing by 37.5% (0.06 m) and 36.4% (0.04 m), respectively. For the 8-h-ahead prediction, the SA-GRU obtains 0.582 NSE (acceptable), 0.087 higher than the 0.495 NSE (unacceptable) of the original GRU, which means the prediction period threshold extends from 7 to 8 h accordingly.
Table 7

NSEs, RMSEs, and MAEs of the original GRU in each prediction period

Prediction period (h)NSERMSE (m)MAE (m)
0.999 0.03 0.01 
0.991 0.05 0.03 
0.952 0.09 0.06 
0.860 0.16 0.11 
0.795 0.19 0.12 
0.685 0.23 0.15 
0.608 0.26 0.18 
0.495 0.29 0.19 
0.277 0.34 0.24 
10 0.212 0.35 0.26 
11 0.096 0.37 0.26 
12 −0.089 0.40 0.28 
Prediction period (h)NSERMSE (m)MAE (m)
0.999 0.03 0.01 
0.991 0.05 0.03 
0.952 0.09 0.06 
0.860 0.16 0.11 
0.795 0.19 0.12 
0.685 0.23 0.15 
0.608 0.26 0.18 
0.495 0.29 0.19 
0.277 0.34 0.24 
10 0.212 0.35 0.26 
11 0.096 0.37 0.26 
12 −0.089 0.40 0.28 
Figure 8

Line chart of NSEs, RMSEs, and MAEs for the original GRU and the SA-GRU on the test set. Optimized values of SA-GRU compared to the original GRU are highlighted in red. Please refer to the online version of this paper to see this figure in colour: https://dx.doi.org/10.2166/hydro.2023.006.

Figure 8

Line chart of NSEs, RMSEs, and MAEs for the original GRU and the SA-GRU on the test set. Optimized values of SA-GRU compared to the original GRU are highlighted in red. Please refer to the online version of this paper to see this figure in colour: https://dx.doi.org/10.2166/hydro.2023.006.

Close modal
Considering that the hyperparameter optimization progress requires massive iterations in which corresponding neural networks with various hyperparameter configurations are generated, it is helpful to analyze the effect of SA by presenting the performance of these networks from the statistical perspective. Figure 9 presents the average and the lowest NSEs of GRU, LSTM, BiLSTM, and BP neural networks generated during iterations in the form of boxplots (the optimal networks are the outputs of the hyperparameter optimization, namely SA-GRU, SA-LSTM, SA-BiLSTM, and SA-BP, respectively). Results show that with the extension of prediction periods, increasing differences between the optimal, the average, and the lowest NSEs of neural networks reflect that all these networks are increasingly affected by the hyperparameter configurations and the optimization effect of SA is increasingly prominent, especially for three RNN neural networks with higher complexity. Notably, the original GRU can hardly achieve the average NSEs of GRU networks generated within the hyperparameter optimization progress, which highlights the great importance of an appropriate hyperparameter configuration.
Figure 9

NSEs of GRU, LSTM, BiLSTM, and BP neural networks generated during iterations in each prediction period. Outliers are removed for clear visualization.

Figure 9

NSEs of GRU, LSTM, BiLSTM, and BP neural networks generated during iterations in each prediction period. Outliers are removed for clear visualization.

Close modal

In summary, by comparing the original GRU and the SA-GRU and exploring the iteration progress of SA itself, the significant effect of the SA-based hyperparameter optimization improves the performance of GRU for urban inundation depth predictions, contributing to not only the increase of precision but also the 1-h extension of the prediction period threshold.

Exploration of the effect of SA on different neural networks

To explore the capacities of different neural networks for urban inundation depth predictions, SA-BP, SA-LSTM, and SA-BiLSTM are developed based on the same dataset. Figure 10 shows the model evaluations in 12 prediction periods and the marked numbers represent the evaluation values obtained by the best models and distinguished by color.
Figure 10

NSE, RMSE, and MAE for SA-BiLSTM, SA-LSTM, SA-GRU, and SA-BP in each prediction period.

Figure 10

NSE, RMSE, and MAE for SA-BiLSTM, SA-LSTM, SA-GRU, and SA-BP in each prediction period.

Close modal

For shorter prediction periods, such as the 1–3-h periods, all three RNN models demonstrate marked performance over the BP neural network, with nearly identical NSE, RMSE, and MAE values. The substantial difference in performance during these early stages of prediction showcases the inherent advantage of RNNs for time series prediction. Examining the 4–6-h prediction periods, the SA-GRU model exhibits a slight edge over the other RNNs, as evidenced by its marginally higher NSE values and lower RMSE and MAE values. However, the distinctions among the RNNs remain subtle, suggesting that each of these RNNs is capable of delivering commendable performance in urban inundation depth prediction. In the context of longer prediction periods, the prediction accuracy of all RNNs demonstrates a similar decline.

Apart from the comparison of prediction accuracy, the training time cost is indeed another crucial factor to consider when selecting the appropriate neural network for urban flood inundation prediction. Figure 11 presents the average training speed (measured in seconds per trial in optimization progress) and the average optimization time (measured in hours per optimization progress) for SA-GRU, SA-LSTM, and SA-BiLSTM.
Figure 11

Comparison between the average training seed and the average optimization time of SA-GRU, SA-LSTM, and SA-BiLSTM.

Figure 11

Comparison between the average training seed and the average optimization time of SA-GRU, SA-LSTM, and SA-BiLSTM.

Close modal

Results show that the SA-GRU has the fastest average training speed, with 139.2 seconds per trial, which is significantly faster than SA-LSTM and SA-BiLSTM, which have average training speeds of 182.5 and 249.5 seconds per trial, respectively. Correspondingly, regarding the average optimization time, the SA-GRU boasts the shortest duration of 3.9 h per optimization progress. In contrast, SA-LSTM and SA-BiLSTM require longer optimization times of 5.1 and 6.9 h, respectively. This difference in training speed means that the SA-GRU model can be trained and optimized more quickly (Siami-Namini et al. 2019; Ozdemir et al. 2022), allowing for more efficient use of resources and time.

In summary, in addition to its slight accuracy advantage over SA-LSTM and SA-BiLSTM, the SA-GRU model is also more time and resource-efficient. Therefore, the SA-GRU could be the preferred choice for urban flood inundation prediction, considering the balance of model performance with computational costs and time constraints.

In this study, we evaluate the efficacy of hyperparameter optimization based on the SA algorithm for predicting urban inundation depth using the GRU. The performance of the proposed SA-GRU model is compared with other benchmark models, including the original GRU, SA-LSTM, SA-BiLSTM, and SA-BP across various prediction periods. The main conclusions of this study can be summarized as follows.

  • 1.

    By comparing the results of prediction periods from 1 to 12, the SA-GRU performs very well for short-term predictions (1–6 h). Since the prediction accuracy of SA-GRU decreases with the extension of prediction periods, the longest prediction period is limited to 8 h: for the 1-h-ahead prediction, NSE, RMSE, and MAE of the proposed model rate obtain 0.999, 0.03, and 0.01, respectively; for the 8-h-ahead limit predictions, NSE, RMSE, and MAE obtain 0.582, 0.26, and 0.17, respectively.

  • 2.

    By comparing the SA-GRU with the original GRU, results show that SA optimization considerably enhances the performance of neural networks. The SA-GRU model consistently outperforms the original GRU model across all prediction periods, which highlights the importance of hyperparameter optimization. This finding contributes to the existing body of research by demonstrating the value of SA optimization in refining neural network performance for urban flood prediction tasks.

  • 3.

    RNN models, including the SA-GRU, outperform the BP neural network in predicting urban inundation depth, which underscores that the inherent advantages of RNNs for the time series prediction tasks. Among the three SA-RNN models (SA-GRU, SA-LSTM, and SA-BiLSTM), the SA-GRU has subtle superiority in terms of prediction accuracy and costs less time for optimization (on average, 3.9 h), rendering it more time and resource-efficient.

  • 4.

    However, there are two main drawbacks worthy of attention. On the one hand, with the extension of prediction periods, uncertainty in the future leads to a decrease in the prediction accuracy of all the models. Therefore, future work could explore alternative optimization techniques, additional features, and data sources to reduce uncertainty and address the challenges in long-term urban inundation depth prediction. On the other hand, the lack of physical explanation is a common problem with black-box models, which affects the adjustability and reliability of the models. Therefore, it is critical to explore and develop physical-informed data-driven models.

In summary, the SA-GRU shows its potential in urban inundation depth prediction with high accuracy for short- to medium-term predictions. However, the limitations for longer-term predictions indicate the need for further research and improvements.

This work was supported by the National Key Research and Development Program of China (Grant No. 2019YFC1510204), the National Natural Science Foundation of China (Grant No. 42175177 and U2240216), and Special Basic Research Key Fund for Central Public Scientific Research Institutes (Grant No. Y521002), the National Natural Science Foundation of China (Grant No. 91847301 and 92047203). The authors would like to thank all experts for their contributions to urban flood protection and control.

Y.Y. performed methodology, format analysis, software, programming, drafting the article, critical reversion. W.Z. conducted methodology and format analysis, wrote the review and edited the manuscript and critical reversion. Y.L. carried out methodology, format analysis, software, data collection, and data curation. Z.L. did format analysis, data analysis, and data curation.

Data cannot be made publicly available; readers should contact the corresponding author for details.

The authors declare there is no conflict.

Abrahart
R. J.
,
Kneale
P. E.
&
See
L. M.
2004
Neural Networks for Hydrological Modelling
.
Balkem, A. A.
CRC Press, London
. https://dx.doi.org/10.1201/9780203024119.
Asadi
S.
,
Shahrabi
J.
,
Abbaszadeh
P.
&
Tabanmehr
S.
2013
A new hybrid artificial neural networks for rainfall-runoff process modeling
.
Neurocomputing
121
,
470
480
.
https://doi.org/10.1016/j.neucom.2013.05.023
.
Avello
E. A.
,
Baesler
F.
&
Moraga
R. J.
2005
A meta-heuristic based on simulated annealing for solving multiple-objective problems in simulation optimization
. In
Proceedings – Winter Simulation Conference
.
https://doi.org/10.1109/WSC.2004.1371355
.
Chang
L. C.
,
Amin
M. Z. M.
,
Yang
S. N.
&
Chang
F. J.
2018
Building ANN-based regional multi-step-ahead flood inundation forecast models
.
Water (Switzerland)
10
,
1
18
.
https://doi.org/10.3390/W10091283
.
Chen
C.
,
Jiang
J.
,
Zhou
Y.
,
Lv
N.
,
Liang
X.
&
Wan
S.
2022
An edge intelligence empowered flooding process prediction using Internet of things in smart city
.
J. Parallel Distrib. Comput.
165
,
66
78
.
https://doi.org/10.1016/J.JPDC.2022.03.010
.
Cho
K.
,
van Merriënboer
B.
,
Bahdanau
D.
&
Bengio
Y.
2014
On the properties of neural machine translation: encoder–decoder approaches
. In:
Proceedings of SSST 2014 – 8th Workshop on Syntax, Semantics and Structure in Statistical Translation
.
Association for Computational Linguistics (ACL)
, pp.
103
111
.
https://doi.org/10.3115/v1/w14-4012
.
Dawson
C. W.
&
Wilby
R.
1998
An artificial neural network approach to rainfall runoff modelling
.
Hydrol. Sci. J.
43
,
47
66
.
https://doi.org/10.1080/02626669809492102
.
de la Fuente
A.
,
Meruane
V.
&
Meruane
C.
2019
Hydrological early warning system based on a deep learning runoff model coupled with a meteorological forecast
.
Water (Switzerland)
11
.
https://doi.org/10.3390/w11091808
.
De Silva
M. M. G. T.
&
Kawasaki
A.
2020
A local-scale analysis to understand differences in socioeconomic factors affecting economic loss due to floods among different communities
.
Int. J. Disaster Risk Reduct.
47
,
101526
.
https://doi.org/10.1016/J.IJDRR.2020.101526
.
Feng
B.
,
Zhang
Y.
&
Bourke
R.
2021
Urbanization impacts on flood risks based on urban growth data and coupled flood models
.
Nat. Hazards
106
,
613
627
.
https://doi.org/10.1007/S11069-020-04480-0/TABLES/3
.
Fu
R.
,
Zhang
Z.
&
Li
L.
2016
Using LSTM and GRU neural network methods for traffic flow prediction
. In:
2016 31st Youth Academic Annual Conference of Chinese Association of Automation (YAC)
. pp.
324
328
.
https://doi.org/10.1109/YAC.2016.7804912
.
Galloway
G. E.
,
Reilly
A.
,
Ryoo
S.
,
Riley
A.
,
Haslam
M.
,
Brody
S.
,
Highfield
W.
,
Gunn
J.
,
Rainey
J.
&
Parker
S.
2018
The Growing Threat of Urban Flooding: 2018
.
Center for Disaster Resilience, University of Maryland. College Park, MD, USA
. pp.
1
44
.
Gao
S.
,
Huang
Y.
,
Zhang
S.
,
Han
J.
,
Wang
G.
,
Zhang
M.
&
Lin
Q.
2020
Short-term runoff prediction with GRU and LSTM networks without requiring time step optimization during sample generation
.
J. Hydrol.
589
,
125188
.
https://doi.org/10.1016/j.jhydrol.2020.125188
.
Gharehbaghi
A.
,
Ghasemlounia
R.
,
Ahmadi
F.
&
Albaji
M.
2022
Groundwater level prediction with meteorologically sensitive Gated Recurrent Unit (GRU) neural networks
.
J. Hydrol.
612
,
128262
.
https://doi.org/10.1016/J.JHYDROL.2022.128262
.
Govt
2020
Global Natural Disaster Assessment Report 2019
.
UN Annu. Rep.
, pp.
1
45
.
Gressling
T.
2020
Automated Machine Learning: Methods, Systems, Challenges
.
Springer, Data Science in Chemistry
.
https://doi.org/10.1515/9783110629453-084
.
Herrera
P. A.
,
Marazuela
M. A.
&
Hofmann
T.
2022
Parameter estimation and uncertainty analysis in hydrological modeling
.
Wiley Interdiscip. Rev. Water
9
,
e1569
.
https://doi.org/10.1002/WAT2.1569
.
Hosseini
F. S.
,
Choubin
B.
,
Mosavi
A.
,
Nabipour
N.
,
Shamshirband
S.
,
Darabi
H.
&
Haghighi
A. T.
2020
Flash-flood hazard assessment using ensembles and Bayesian-based machine learning models: application of the simulated annealing feature selection method
.
Sci. Total Environ.
711
,
135161
.
https://doi.org/10.1016/J.SCITOTENV.2019.135161
.
Hsu
K.
,
Gupta
H. V.
&
Sorooshian
S.
1995
Artificial neural network modeling of the rainfall-runoff process
.
Water Resour. Res.
31
,
2517
2530
.
Jansen
T.
2011
Simulated annealing
. In:
Theory of Randomized Search Heuristics: Foundations and Recent Developments
, pp.
171
195
.
https://doi.org/10.1142/9789814282673-6
.
Kan
G.
,
Yao
C.
,
Li
Q.
,
Li
Z.
,
Yu
Z.
,
Liu
Z.
,
Ding
L.
,
He
X.
&
Liang
K.
2015
Improving event-based rainfall-runoff simulation using an ensemble artificial neural network based hybrid data-driven model
.
Stoch. Environ. Res. Risk Assess.
29
,
1345
1370
.
https://doi.org/10.1007/s00477-015-1040-6
.
Kao
I. F.
,
Zhou
Y.
,
Chang
L. C.
&
Chang
F. J.
2020
Exploring a long short-term memory based encoder-decoder framework for multi-step-ahead flood forecasting
.
J. Hydrol.
583
,
124631
.
https://doi.org/10.1016/J.JHYDROL.2020.124631
.
Kao
I. F.
,
Liou
J. Y.
,
Lee
M. H.
&
Chang
F. J.
2021
Fusing stacked autoencoder and long short-term memory for regional multistep-ahead flood inundation forecasts
.
J. Hydrol.
598
,
126371
.
https://doi.org/10.1016/j.jhydrol.2021.126371
.
Kirkpatrick
S.
,
Gelatt
C. D.
&
Vecchi
M. P.
1983
Optimization by simulated annealing
.
Science
220
,
671
680
.
https://doi.org/10.1126/science.220.4598.671
.
Kratzert
F.
,
Klotz
D.
,
Brenner
C.
,
Schulz
K.
&
Herrnegger
M.
2018
Rainfall-runoff modelling using Long Short-Term Memory (LSTM) networks
.
Hydrol. Earth Syst. Sci.
22
,
6005
6022
.
https://doi.org/10.5194/hess-22-6005-2018
.
Kuhn
M.
&
Johnson
K.
2013
Applied Predictive Modeling
.
https://doi.org/10.1007/978-1-4614-6849-3
.
Kumar
D.
,
Singh
A.
,
Samui
P.
&
Jha
R. K.
2019
Forecasting monthly precipitation using sequential modelling
.
Hydrol. Sci. J.
64
,
690
700
.
https://doi.org/10.1080/02626667.2019.1595624
.
Li
W.
,
Kiaghadi
A.
&
Dawson
C.
2021
High temporal resolution rainfall–runoff modeling using long-short-term-memory (LSTM) networks
.
Neural Comput. Appl.
33
,
1261
1278
.
https://doi.org/10.1007/s00521-020-05010-6
.
Luo
P.
,
Peng
J.
,
Guan
Z.
&
Fan
J.
2018
Dual regularized multi-view non-negative matrix factorization for clustering
.
Neurocomputing
294
,
1
11
.
https://doi.org/10.1016/J.NEUCOM.2017.10.023
.
Madhuri
R.
,
Raja
Y. S. L. S.
,
Raju
K. S.
,
Punith
B. S.
&
Manoj
K.
2021
Urban flood risk analysis of buildings using HEC-RAS 2D in climate change framework
.
H2Open J.
4
,
262
275
.
https://doi.org/10.2166/H2OJ.2021.111
.
Mei
P.
,
Li
M.
,
Zhang
Q.
,
Li
G.
&
Song
L.
2022
Prediction model of drinking water source quality with potential industrial-agricultural pollution based on CNN-GRU-Attention
.
J. Hydrol.
610
,
127934
.
https://doi.org/10.1016/J.JHYDROL.2022.127934
.
Metropolis
N.
,
Rosenbluth
A. W.
,
Rosenbluth
M. N.
,
Teller
A. H.
&
Teller
E.
1953
Equation of state calculations by fast computing machines
.
J. Chem. Phys.
21
,
1087
1092
.
Nguyen
D. H.
&
Bae
D.-H.
2020
Correcting mean areal precipitation forecasts to improve urban flooding predictions by using long short-term memory network
.
J. Hydrol.
584
,
124710
.
https://doi.org/10.1016/j.jhydrol.2020.124710
.
Ozdemir
A. C.
,
Buluş
K.
&
Zor
K.
2022
Medium- to long-term nickel price forecasting using LSTM and GRU networks
.
Resour. Policy
78
,
102906
.
https://doi.org/10.1016/J.RESOURPOL.2022.102906
.
Rangari
V. A.
,
Umamahesh
N. V.
&
Bhatt
C. M.
2019
Assessment of inundation risk in urban floods using HEC RAS 2D
.
Model. Earth Syst. Environ.
54
(
5
),
1839
1851
.
https://doi.org/10.1007/S40808-019-00641-8
.
Roy
S.
,
Banerjee
R.
&
Bose
P. K.
2014
Performance and exhaust emissions prediction of a CRDI assisted single cylinder diesel engine coupled with EGR using artificial neural network
.
Appl. Energy
119
,
330
340
.
https://doi.org/10.1016/J.APENERGY.2014.01.044
.
Rumelhart
D.
,
Hinton
G.
&
Williams
R.
1986
Learning internal representations by error propagation (original)
. In:
Explorations in the Microstructure of Cognition
, Vol.
1
.
California Univ San Diego La Jolla Inst for Cognitive Science
.
Sahoo
S. N.
&
Sreeja
P.
2016
Relationship between peak rainfall intensity (PRI) and maximum flood depth (MFD) in an urban catchment of Northeast India
.
Nat. Hazards
83
,
1527
1544
.
https://doi.org/10.1007/S11069-016-2374-1/FIGURES/13
.
Siami-Namini
S.
,
Tavakoli
N.
&
Namin
A. S.
2019
The performance of LSTM and BiLSTM in forecasting time series
. In:
Proc. – 2019 IEEE Int. Conf. Big Data, Big Data 2019
, pp.
3285
3292
.
https://doi.org/10.1109/BigData47090.2019.9005997
.
Soon
F. C.
,
Khaw
H. Y.
,
Chuah
J. H.
&
Kanesan
J.
2018
Hyper-parameters optimisation of deep CNN architecture for vehicle logo recognition
.
IET Intell. Transp. Syst.
12
,
939
946
.
https://doi.org/10.1049/iet-its.2018.5127
.
Tamiru
H.
&
Dinka
M. O.
2021
Application of ANN and HEC-RAS model for flood inundation mapping in lower Baro Akobo River Basin, Ethiopia
.
J. Hydrol. Reg. Stud.
36
,
100855
.
https://doi.org/10.1016/j.ejrh.2021.100855
.
Valeh
S.
,
Motamedvairi
B.
,
Kiadaliri
H.
&
Ahmadi
H.
2021
Hydrological simulation of Ammameh basin by artificial neural network and SWAT models
.
Phys. Chem. Earth, Parts A/B/C
123
,
103014
.
https://doi.org/10.1016/J.PCE.2021.103014
.
Wang
B.
,
Loo
B. P. Y.
,
Zhen
F.
&
Xi
G.
2020
Urban resilience from the lens of social media data: responses to urban flooding in Nanjing, China
.
Cities
106
,
102884
.
https://doi.org/10.1016/j.cities.2020.102884
.
Wang
J.
,
Mu
X.
,
Chen
S.
,
Liu
W.
,
Wang
Z.
&
Dong
Z.
2021
Dynamic evaluation of water resources carrying capacity of the Dianchi Lake Basin in 2005–2015, based on DSPERM framework model and simulated annealing-projection pursuit model
.
Reg. Sustain.
2
,
189
201
.
https://doi.org/10.1016/J.REGSUS.2021.06.003
.
Wang
P.
,
Li
Y.
,
Yu
P.
&
Zhang
Y.
2021
The analysis of urban flood risk propagation based on the modified susceptible infected recovered model
.
J. Hydrol.
603
,
127121
.
https://doi.org/10.1016/j.jhydrol.2021.127121
.
Wang
J.
,
Li
X.
,
Li
J.
,
Sun
Q.
&
Wang
H.
2022
NGCU: a new RNN model for time-series data prediction
.
Big Data Res.
27
,
100296
.
https://doi.org/10.1016/j.bdr.2021.100296
.
Wang
Z.
,
Tian
J.
&
Feng
K.
2022
Optimal allocation of regional water resources based on simulated annealing particle swarm optimization algorithm
.
Energy Rep.
8
,
9119
9126
.
https://doi.org/10.1016/J.EGYR.2022.07.033
.
Waseem
M.
,
Mani
N.
,
Andiego
G.
&
Usman
M.
2008
A review of criteria of fit for hydrological models
.
Int. Res. J. Eng. Technol.
9001
,
1765
.
Widiasari
I. R.
,
Nugoho
L. E.
,
Widyawan
&
Efendi
R.
2018
Context-based hydrology time series data for a flood prediction model using LSTM
. In:
Proceedings – 2018 5th International Conference on Information Technology, Computer and Electrical Engineering, ICITACEE 2018
.
https://doi.org/10.1109/ICITACEE.2018.8576900
.
Xiang
Z.
,
Yan
J.
&
Demir
I.
2020
A rainfall-runoff model with LSTM-based sequence-to-sequence learning
.
Water Resour. Res.
56
.
https://doi.org/10.1029/2019WR025326
.
Xu
Y.
,
Hu
C.
,
Wu
Q.
,
Jian
S.
,
Li
Z.
,
Chen
Y.
,
Zhang
G.
,
Zhang
Z.
&
Wang
S.
2022
Research on particle swarm optimization in LSTM neural networks for rainfall-runoff simulation
.
J. Hydrol.
608
,
127553
.
https://doi.org/10.1016/j.jhydrol.2022.127553
.
Yang
L.
&
Shami
A.
2020
On hyperparameter optimization of machine learning algorithms: theory and practice
.
Neurocomputing
415
,
295
316
.
https://doi.org/10.1016/j.neucom.2020.07.061
.
Yi
H.
&
Yang
J.
2014
Research trends of post disaster reconstruction: the past and the future
.
Habitat Int.
42
,
21
29
.
https://doi.org/10.1016/J.HABITATINT.2013.10.005
.
Yuan
X.
,
Chen
C.
,
Lei
X.
,
Yuan
Y.
&
Muhammad Adnan
R.
2018
Monthly runoff forecasting based on LSTM–ALO model
.
Stoch. Environ. Res. Risk Assess.
32
,
2199
2212
.
https://doi.org/10.1007/s00477-018-1560-y
.
Zhang
J.
,
Zhu
Y.
,
Zhang
X.
,
Ye
M.
&
Yang
J.
2018
Developing a long short-term memory (LSTM) based model for predicting water table depth in agricultural areas
.
J. Hydrol.
561
,
918
929
.
https://doi.org/10.1016/J.JHYDROL.2018.04.065
.
Zuo
G.
,
Luo
J.
,
Wang
N.
,
Lian
Y.
&
He
X.
2020
Decomposition ensemble model based on variational mode decomposition and long short-term memory for streamflow forecasting
.
J. Hydrol.
585
,
124776
.
https://doi.org/10.1016/J.JHYDROL.2020.124776
.
This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (CC BY 4.0), which permits copying, adaptation and redistribution, provided the original work is properly cited (http://creativecommons.org/licenses/by/4.0/).