Floods are becoming increasingly frequent and severe due to climate change and urbanization, thereby increasing risks to lives, property, and the environment. This necessitates the development of precise flood forecasting systems. This study addresses the critical task of predicting flood peak arrival times, which is essential for timely warnings and preparations, by introducing a comprehensive machine-learning framework. Our approach integrates interpretable feature engineering, individual model design, and novel model ensembles to enhance prediction accuracy. We extract informative features from historical flood flow and rainfall data, design a suite of machine-learning models, and develop a novel ensemble technique to combine model predictions. We conducted case studies on the Tunxi and Changhua basins in China. Numerical experiments reveal that our method significantly benefits from feature engineering and model ensembles, achieving mean absolute error (MAE) prediction errors of 1.524 h for Tunxi and 2.192 h for Changhua. These results notably outperform the best baseline method, which achieves MAE errors of 1.727 h for Tunxi and 2.737 h for Changhua.

  • This paper identifies and formalizes an important flood forecasting task – predicting the arrival time of flood peak.

  • This paper develops a novel machine-learning prediction framework, built based on an interpretable feature engineering and model ensemble method.

  • This paper demonstrates the importance and effectiveness of feature engineering and ensemble methods in improving prediction accuracy.

Floods are among the most frequent and impactful natural disasters. With their increasing frequency and intensity, driven by climate change and urbanization, they present significant risks to human life, property, and environmental health. Accurate flood monitoring and forecasting are imperative for fast reaction and early intervention, enabling timely evacuation plans and resource allocation to mitigate potential damages. Additionally, precise flood predictions empower decision-makers to implement proactive measures, such as deploying temporary flood barriers or issuing warnings to vulnerable communities, ultimately reducing the socioeconomic impacts of flooding events (Jia et al. 2018; Lee & Kim 2021; Li et al. 2021).

In flood prevention and mitigation, flood forecasting plays a pivotal role in shaping effective strategies (Badrzadeh et al. 2015; Yenming et al. 2016; Zhihui et al. 2023). Along this line of research, a category of works relies on conventional hydrological models (Deng & Li 2013; Han & Coulibaly 2017; Le et al. 2019; Xu et al. 2019; Zhao et al. 2019), such as the Xin'an Jiang model, which simulates standard flood processes using established physical principles. More recently, as the model-based forecasting methods heavily rely on human expertise and may be challenging to adapt to more general settings, a more popular and practical option is to develop data-driven models, i.e., building models upon extensive observations and parameterized frameworks (Garg et al. 2023; Rajab et al. 2023; Sharma & Kumari 2024).

In particular, the data-driven model typically leverages historical data, especially rainfall and flood flow data, to make predictions for future times. The critical part is the forecasting model. A number of recent studies attempted to leverage machine-learning algorithms to perform the forecasting (Toth et al. 2000; Hapuarachchi et al. 2011; Palash et al. 2018; Ming et al. 2020; Nguyen & Chen 2020; Ashok & Pekkat 2022; Kabbilawsh et al. 2022; Rahman et al. 2023). These methods encompass conventional machine-learning models such as linear models (Palash et al. 2018) and multi-feature linear models (Rahman et al. 2023), as well as deep neural networks such as long short-term memory (LSTM) (Le et al. 2019; Cho et al. 2022; Atashi et al. 2022; Rajab et al. 2023; Hayder et al. 2023), wavelet-based neural network models (Agarwal et al. 2022; Santos et al. 2023; Chakraborty & Biswas 2023), and spatial–temporal models (Ding et al. 2020; Noor et al. 2022). Moreover, some researchers have further adopted these models to perform more complicated tasks with improved modifications. To name a few, Agarwal et al. (2023) have integrated a class of machine-learning models to study the extreme annual rainfall in North-Eastern India. The same group has also extended the utilization of neural network models to the multiple input multiple output (MIMO) setting to perform the flood modeling and learn storage rate changes in the river system (Agarwal et al. 2021a, 2021b). These models have been demonstrated to be effective in performing short-term flood prediction.

However, rather than performing short-term flood flow prediction, predicting the arrival time of the flood peak – or, more precisely, the time when the flood flow reaches a dangerous level – could be even more crucial. This is because it enables the government to carry out early warnings and make necessary intervention efforts and preparations. Significantly, flood peak arrival time prediction is generally more challenging than short-term flood flow prediction as it typically requires the model to anticipate flood behavior over a longer term. That is, the flood peak could occur much later than the prediction time. Therefore, the machine-learning methods developed in the previous literature may not be directly adapted to perform accurate flood peak arrival time prediction. To this end, the objectives of this paper are summarized as follows:

  • We will develop new machine-learning frameworks for predicting the flood peak arrival times based on the historical data.

  • We will explore a set of modules in the developed machine-learning framework that can enable more accurate, robust, and interpretable predictions.

  • We will conduct case studies in the Tunxi and Changhua basins to validate our framework's design and compare its prediction accuracy with baseline methods.

The remaining part of this paper is organized as follows: First, we will first introduce the primary method in Section 2, covering the problem overview and description of the developed machine-learning framework. This encompasses feature engineering, machine-learning model design, and model ensemble.

In Section 3, we will introduce the case studies in Tunxi and Changhua, China, including the description of the dataset and some basic data organization and preprocessing procedures. In Section 4, we will provide a detailed illustration of the performance of the proposed models on several real-world flood forecasting datasets. This section includes a comparison between the proposed forecasting method and baselines, as well as ablation studies to interpret the role of different features and modules in our method. In Section 5, we discuss the key findings and the limitations of the developed approaches. Then, we summarize the methodology and results of this paper, as well as highlight the future directions in Section 6.

Overview

As we mentioned previously, most existing models focus on developing models for predictions in the fixed forecast horizons (1, 6, 12, 24, 36, 48 h). Instead, this paper's objective is to predict the arrival time of the flood peak at the current moment using all available information. In the problem constructed in this paper, we can use data from the current moment and all historical data up to the current moment, including historical flow data from various flow stations and historical rainfall data from various rain stations. Mathematically, this can be represented as:
(1)

More specifically, since the time of making a prediction is known, the target of the prediction can be transformed into the time difference between the current one and the flood peak arrival time, i.e., the time to the flood peak. This will serve as the prediction target in the model design part.

To effectively predict the timing of flood peaks using historical flow and rainfall data, several challenges must be addressed. First, there is no direct physical model that links the timing of a flood peak to historical patterns. Our understanding relies primarily on general observations, such as heavier rainfall typically shortening the lead time to a flood peak, higher flow indicating more intense flood peaks, and prolonged heavy rain leading to quicker rises in flood levels. Second, identifying the most appropriate machine-learning model remains uncertain, as it is unclear which model will provide the best prediction performance. Consequently, we are motivated to investigate the design of both feature extraction methods and predictive models, which are detailed in the subsequent subsections.

Feature engineering

In this part, we aim to develop a simple yet effective method for extracting informative features from the raw data. Such a feature engineering framework has been widely demonstrated to be effective in improving the performance of the model in many time series-based machine-learning tasks (Wang et al. 2022; Pablo & Dormido-Canto 2023).

In particular, we will use the historical rainfall and flood flow as the raw features and then design a novel approach to extract the features accordingly. We use feature engineering to account for the duration, amount, and frequency of events. For example, average rainfall and river flow measure the strength of these factors in the area. Their variability indicates the potential for sudden, intense rain, and how often it rains consecutively suggests how frequently the area experiences rainfall leading up to peak floods. We detail these features, including their mathematical formula and descriptions, as follows.

Firstly, let be the rainfall data at the historical timestamps and be the daily flood flow increase at the historical timestamps. Our goal is to design a collection of interpretable features for predicting the future arrival time of the flow peak at time t. We will then perform the feature generation approaches for the historical rainfall and flood flow separately. In particular, to avoid very long historical sequences, we consider a fixed window size L and only make use of the L most recent data, i.e., and . The feature engineering details will be introduced below and summarized in Figure 1. Here, we aim to extract more informative features from the raw historical flood and rainfall data from the perspectives of frequency, count of heavy rain/flood, statistical quantities, and increases/changes.
Figure 1

Visualization of the feature engineering module, where we extract informative features from the raw historical flood and rainfall data through the perspective of frequency, count of heavy rain/flood, statistical quantities, and increases/changes.

Figure 1

Visualization of the feature engineering module, where we extract informative features from the raw historical flood and rainfall data through the perspective of frequency, count of heavy rain/flood, statistical quantities, and increases/changes.

Close modal

Features of historical rainfall

We first consider the features generated by leveraging the historical rainfall. In particular, the corresponding candidate features and their calculation formula are described as follows:

  • Cumulative rainfall, formulated as . The cumulative rainfall basically calculates the total rainfall of the historical period. In general, higher cumulative rainfall typically implies more heavier floods and thus plays an important role in predicting the peak arrival time.

  • Average rainfall, formulated as . Here, one may think the average rainfall is similar to the cumulative rainfall since there is only a simple linear transformation between them. However, in our calculation, we consider the timestamps with nonzero rainfall (i.e., ri > 0) and only take the average over the nonzero rainfalls. Notably, the number of rainy days in the L's most recent timestamps is different for different timestamps. This implies that the average rainfall considered in our work is, in fact, a nonlinear function of the cumulative rainfall and thus can help provide richer information for our prediction task.

  • Exponential moving average of rainfall, formulated as . Here ∈ (0, 1) denotes the decaying factor, which is chosen to be 2/3 in our paper. In particular, we consider the exponential moving average because it is natural that the rainfall in more recent times will have a stronger effect on the prediction, so we increase the weights for the closer rainfall data.

  • Standard deviation of daily rainfall, formulated as . The idea of considering the standard deviation of the rainfall is to investigate whether the stability of the rainfall will affect the flood prediction. For instance, if we have the same total rainfall, which of the following two cases is more likely to reach the flood peak earlier: (1) there is a very heavy rainfall in one single timestamp; (2) the rainfall is uniformly distributed across all timestamps.

  • Rainfall increase, formulated as . Here we use 5-day average to smooth the calculation. In general, the idea of calculating the rainfall increase is to show whether there is heavier rainfall in recent days, making it more likely to have heavier floods in the future.

  • Count of heavy rainfall, formulated as , where is a threshold that can be set according to the definition in (MANOBS 2013), and the rainfall higher than is considered as the heavy rainfall. We apply this feature as a complement to the total/average rainfall, which can be understood as the frequency of heavy rainfall rather than just the magnitude of rainfall.

  • Count of consecutive heavy rainfall, formulated as , the number of consecutive days with heavy rainfall is comparable to the overall count of heavy rainfall. Nevertheless, there are cases where it is more informative to determine if heavy rainfall occurs on successive days. This intuitively provides a better understanding of the frequency and intensity of the rainfall.

Features of historical flood flow

Then, we consider the features generated by leveraging the historical flood flow. Compared to the features of historical rainfall, the feature extraction method will be similar. We consider the strength, frequency, and standard deviation of the daily changes in flood flow. In particular, the features and the corresponding formulas are provided as follows:

  • Previous flood flow , i.e., the flood flow in the previous time stamp.

  • Average flood flow change, formulated as .

  • Exponential moving average, formulated as .

  • Standard deviation, formulated as .

  • Flow increase, formulated as: .

Now we have detailed a set of analytic and interpretable feature extraction methods. In the experiment section (Section 3), we will first apply a standard statistical method to evaluate the importance of these features and show that using such interpretable feature extraction methods can lead to substantially better predictive performance compared to raw features.

Predictive model designs

In this section, we will introduce the predictive models used for forecasting the arrival time of flood peaks. Given the uncertainty regarding which machine-learning model is optimal for this task, we will consider a set of standard machine-learning models as well as several task-specific models. Instead of selecting a single best model based on the smallest validation error, we will construct an ensemble model. This ensemble approach utilizes all candidate models by assigning them different importance weights, as described in the next subsection.

In particular, each model will take the features generated by the above-mentioned feature engineering and generate the time to the flood peak arrival time. Mathematically, let t be the current time, then based on Section 2.2, we can obtain a collection of the features, denoted by , by performing calculations on the raw data, i.e., recent historical rainfall and flood flow within a fixed window size. Then let be the flood peak arrival time, the machine-learning model aims to take the features as input and perform the following prediction on the time to the flood peak, i.e., :
(2)
where denotes the model prediction and denotes the target prediction.

Moreover, we will consider the following machine-learning models that will be used for the model ensemble.

Multi-factor linear regression The multi-factor linear regression model investigates the regression between one dependent variable and multiple explanatory variables and plays a significant role in various fields, such as economics (where it has led to the development of multi-factor stock selection models) and biostatistics (Tsakiri et al. 2018; Mohammad et al. 2020; James et al. 2023). Multi-factor linear regression simulates the linear relationship between one dependent variable and several explanatory variables. Mathematically, let be K features generated from our feature engineering approach. The prediction result can be formulated using the following equation:
(3)
where denotes the slope parameters for K features separately and b denotes the intercept parameter. Then, considering the i-th data point, let and be the ground truth and the prediction result of the target data. Denote Xi as the features, linear regression seeks to minimize the following objective function:
(4)

Then, after obtaining the parameter and b, we are able to use Equation (3) to output the prediction, i.e., the time to the flood peak.

Indeed, the coefficients of multi-factor linear regression offer higher interpretability than other black-box models, such as neural networks. In multi-factor linear regression, each coefficient represents the average impact of an independent variable on the dependent variable, assuming other variables remain constant. This relationship is easier to understand and explain. Therefore, multi-factor linear regression is preferred for its greater interpretability in serious contexts. However, one critical issue in linear regression is its multicollinearity, i.e., some features are correlated with each other, leading to model redundancy. In order to address such a potential issue, we further consider two regularized linear regression methods: ridge regression (Gary 2009; Carlos & Upmanu 2010; Mumtaz et al. 2020) and Lasso regression (Ranstam & Cook 2018; Rastogi et al. 2021; Rofah et al. 2021).

Ridge regression: One method to mitigate the feature redundancy, or more precisely, the overfitting phenomenon caused by the feature redundancy, is to penalize the magnitudes of the linear model parameters . Ridge regression leverages the ℓ2 norm over the weights as the penalization, which is added to the original objective function (3), leading to the following regularized objective:
(5)
where is a tunable positive parameter to adjust the trade-off between the model performance and redundant feature utilization.
Lasso regression: In practice, while ridge regression incorporates regularization to mitigate overfitting, its primary focus is not specifically on addressing redundant features. For a more explicit handling of redundant features, it is necessary to limit the number of effective weights by assigning a zero weight to redundant features. This concept aligns precisely with the principle of Lasso regression, which seeks a sparse weight vector through an ℓ1 norm penalty. Mathematically, the objective of Lasso regression is to minimize the following loss:
(6)
where is a tunable positive parameter for Lasso regression.
Multi-layer perceptron (MLP): The neural network model (ANN), also known as MLP, has been widely adopted in many areas to deal with complicated data (Taud & Mas 2018; Suddul et al. 2020; Haribabu et al. 2021; Pedro et al. 2021) with nonlinear correlations. It has lower interpretability than linear models but typically can achieve better performance. In particular, it can be regarded as stacking multiple layers on the input data, where each layer performs a linear operation on the input followed by a nonlinear activation function. In our task, we consider a four-layer model, which can be formulated as:
(7)
where and are the model parameters of the MLP, X denotes the collection of all features, denotes the predicted value, and denotes the rectified linear unit (ReLU) activation. Then, based on n training data points, the MLP weights are obtained by training with the following objective function:
(8)
Feature-aware MLP. The standard MLP model applies a linear operation directly to the feature vector, failing to distinguish between rainfall and flood features. However, we believe that these two sets of features may represent different aspects of flooding and thus should be processed separately before being combined for prediction. To address this, we propose a novel model called the feature-aware MLP. This model first applies separate sub-blocks to the extracted rainfall features and flood flow features, resulting in two distinct embeddings. These embeddings are then merged using an additional MLP sub-block, which ultimately outputs the prediction. In particular, we have three MLP sub-blocks, denoted by MLP1, MLP2, and MLP3. Let and be the extracted flood and rainfall feature vectors, respectively. Then, the prediction of the feature-aware MLP can be obtained as follows:
(9)
Here we stack the outputs of MLP1 and MLP2 and then feed them into MLP3 to obtain the prediction. The visualization of the feature-aware MLP is displayed in Figure 2. Then, the model training is the same as that for MLP, where the mean square loss is used as the training objective.
Figure 2

Visualization of the feature-aware MLP model, outputs of MLP1 and MLP2, which operated on flood and rainfall features separately, and then feeding them into MLP3 to obtain the prediction.

Figure 2

Visualization of the feature-aware MLP model, outputs of MLP1 and MLP2, which operated on flood and rainfall features separately, and then feeding them into MLP3 to obtain the prediction.

Close modal

Model ensemble

In the previous section, we introduced multiple machine-learning models applicable to our task. However, their performance can vary significantly across different datasets, making it uncertain which model will achieve the best results. To address this, inspired by recent advancements in ensemble models (Wu et al. 2014; Shijun et al. 2020; Bracher et al. 2021; Zhai et al. 2022; Kim et al. 2023), we propose combining the outputs of these models through ensembling. Specifically, our model ensemble will adhere to two principles: (1) the method should be lightweight, avoiding modifications to already trained models, and (2) we should fully utilize the validation dataset to design the most effective ensembling method. To this end, we consider a simply linear combination of the prediction results of the aforementioned models, which can be seen as a fine-tuning based on a set of rough estimations. In particular, given an input data X, let be the corresponding predictions generated by the aforementioned five models. The model ensemble follows the following procedure below:
(10)
where are the coefficients for these models. Then, since these base models have already been well trained on the training data, we will not readjust these weights based on the training data. Instead, we will utilize the validation dataset for a fair adjustment. In particular, considering M validation data points , these weights are optimized by solving the following mean square error (MSE) loss:
(11)
where are the predictions with respect to by all base models, respectively.

Model pipeline and workflow

Based on the aforementioned parts, the entire pipeline of our machine-learning forecasting framework is summarized in Figure 3, which combines feature extraction, multiple model inferences, and model ensemble. They are organized in a sequential manner. More specifically, the workflow of the entire framework development is detailed as follows:
  • Step 1: Perform feature engineering on the raw historical data. Then, a linear test of the features obtained from feature engineering will be applied to perform feature cleaning.

  • Step 2: Based on the extracted features, we train multiple models to generate the candidate predictions. These models are trained based on the MSE loss with possible regularization terms. The hyperparameters are tuned through a grid search method.

  • Step 3: Given the five trained candidate models, we further implement a linear ensemble for their predictions. The coefficients are trained and optimized based on the MSE loss on the validation dataset.

  • Step 4: Once all model parameters are well trained, given a new set of raw features, we will then do feature extraction, candidate model inferences, and results ensemble sequentially to obtain the final prediction.

Figure 3

Visualization of the machine-learning forecasting framework, including feature extraction, model design, and model ensemble.

Figure 3

Visualization of the machine-learning forecasting framework, including feature extraction, model design, and model ensemble.

Close modal

Model evaluation

Using multiple evaluation metrics is crucial for assessing model performance from different viewpoints. This comprehensive approach guarantees that we capture various aspects of the model's accuracy and robustness, providing a more complete picture of its effectiveness. A method can be considered strictly superior if it surpasses other methods across all these metrics.

In our work, we will utilize several key metrics: rooted mean squared error (RMSE), mean absolute error (MAE), and coefficient of determination (R2). Specifically, RMSE measures the average squared difference between predicted and actual values. MAE calculates the average absolute difference between predicted and actual values. R2 indicates the proportion of the variance in the dependent variable that can be predicted from the independent variables. The formulas of these metrics are given as follows:
(12)
(13)
(14)
where denotes the predicted value for the i-th data, denotes the i-th ground truth target, n denotes the total sample size, and denotes the mean of ground truth targets. In this entire paper, the RMSE and MAE are evaluated in hours.
In this section, we will briefly introduce the case studies using real-world flood data from certain areas. In particular, we consider two typical basins: Tunxi and Changhua. The Tunxi Basin is located in Tunxi District, Huangshan City, Anhui Province, China. It contains 11 rain gauge stations and 1 flow station, and the station distribution is shown in Figure 4(a). The Changhua Basin is located in Lin'an District, Hangzhou City, Zhejiang Province, China. It contains seven rain gauge stations and one flow station, and the station distribution is shown in Figure 4(b).
Figure 4

(a) Illustration of the Tunxi Basin, China, (b) Illustration of the Changhua Basin, China.

Figure 4

(a) Illustration of the Tunxi Basin, China, (b) Illustration of the Changhua Basin, China.

Close modal

This section is then organized as follows: first, we will introduce the key properties of the dataset in these two basins. Then, we will evaluate the feature engineering module in our framework and discuss the importance of different features. Finally, we will compare the performances achieved by different methods.

Description of the dataset

In total, the Tunxi Basin dataset contains 134 floods, including 23,132 historical monitoring records (rainfall and flow) in total, spanning from 17 June 1981 to 18 March 2007. The Changhua Basin dataset contains 44 floods, including 9,354 historical monitoring records (rainfall and flow) in total, spanning from 7 April 1998 to 20 July 2010. The detailed description of the dataset is shown in Table 1. The dataset forms a time series, and the features contain the rainfall and cumulative flood flow at each timestamp, the granularity of the time series data is 1 h.

Table 1

Description of the Tunxi and Changhua dataset

DatasetTunxi Basin datasetChanghua Basin dataset
Type of data Rainfall and flow Rainfall and flow 
Sampling frequency 1 data/h 1 data/h 
Number of floods 134 floods 44 floods 
Number of data in the dataset 23,132 9,354 
DatasetTunxi Basin datasetChanghua Basin dataset
Type of data Rainfall and flow Rainfall and flow 
Sampling frequency 1 data/h 1 data/h 
Number of floods 134 floods 44 floods 
Number of data in the dataset 23,132 9,354 

Moreover, we display all floods’ data in Figure 5, where each curve represents the time series for one single flood period, and the x-axis denotes the relative timestamps of the flow (the unit is hour). In Figure 5(a), it can be clearly seen that the majority of floods in the Tunxi Basin have a duration of around 150 h, with the peak generally occurring around 50 h.
Figure 5

Visualization of all floods in the Tunxi and Changhua basins, China. (a) TunXi Basin, (b) Changhua Basin.

Figure 5

Visualization of all floods in the Tunxi and Changhua basins, China. (a) TunXi Basin, (b) Changhua Basin.

Close modal

Data organization

Firstly, as the original data are in the form of a time series, we will first reshape it into a data array. In particular, let be the original data. We set the window size as h, i.e., the number of historical data we use for prediction. Then the features of our dataset will be formulated as:
(15)

Then let , the prediction target, i.e., the time to the flood peak, is defined as , where denotes the time of the closest flood peak. Both and are described in hours. This leads to a feature–target pair . Moreover, since it is difficult to predict the flood peak at a very early stage (e.g., ), we will only select the data for which the ground truth time to the flood peak is less than 20 h. This gives us the dataset with raw features.

In this section, we will present the numerical results for the case studies in the Tunxi and Changhua basins. This section is organized as follows: first, we will introduce and evaluate the feature engineering module in our framework. Then, we discuss the importance of different features. Finally, we will make comparisons between the performances achieved by different methods to demonstrate the benefit of the proposed framework.

Feature engineering

Preprocessing

Based on the data described in Section 3.2 and the methods described in Section 2.2.2, we then perform feature engineering. Multiple informative features are extracted from the raw data and then added as inputs to the model. These features include two main categories: rainfall and flow, totaling 13 features. Due to the significant numerical differences contained in hydrological data, it is necessary to preprocess the features before model training.

  • Perform min–max normalization on the input variables to standardize the data to a range of 0–1, reducing the impact of data representation methods on the model results.

  • Shuffle the order of the dataset. By shuffling the order of the dataset, it prevents the occurrence of patterned data, which could lead to overfitting or non-convergence.

  • Perform train-validation-test splitting using a 6:2:2 ratio.

Analysis of the features

In response to the designed features, we first employ the single-factor regression to evaluate the performance of these features. Based on this, we can demonstrate which of them are effective. We provide the results for the Tunxi and Changhua datasets in Table 2, where the p-value, RMSE, MAE, and on the test dataset are presented.

Table 2

Analysis of the features for Tunxi Basin, where the important features with are marked in green, and the irrelevant features with p-value > 0.05 are marked in magenta.

 
 

RMSE and MAE are measured in hours.

From the results, it can be seen that the p-values for the daily increase are consistently insignificant in both Changhua and Tunxi. This implies that this feature should be excluded in the feature integration section below. Additionally, the accumulated rainfall, the days of rainfall, and the number of days of consecutive heavy rain all achieved an of around 0.5 in out-of-sample tests, proving that these features have a specific effect on flood peak prediction. The exponential weighted moving average (EWMA) of rainfall performs slightly better than the average rainfall, indicating that recent rainfall should indeed have a more significant impact on the flood peak. This aligns with logical common sense, as closer rainfall events are more influential. However, when examining the flow features, although the p-values of each feature are significant, their performance is substantially worse. This suggests that while we are aiming for predictions related to flood flow, relying solely on flow features does not yield good performance. It seems that features derived from rainfall data may be more crucial for our task.

Feature redundancy analysis

In the previous section, we observed the advantages of using a multi-factor model to predict the arrival time of flood peaks, demonstrating its feasibility for this task. However, when dealing with multi-factor models, feature redundancy needs to be considered. We perform principal component analysis (PCA) (Hervé & Williams 2010) on all features to analyze the feature redundancy. In particular, we use the percentage of variance explained to investigate the effective dimension of the feature space:
(16)
where denotes the sorted eigenvalues of the covariance matrix of training data points in descending order, d denotes the number of model parameters, and k denotes the dimensionality of the new feature space. The results are displayed in Figure 6(a) and 6(b). It is evident that, for these two datasets, eight features are sufficient to explain nearly the entire feature space. Therefore, in the following experiments, we will remove the irrelevant and redundant features and only keep the remaining features to perform the flood peak arrival time forecasting.
Figure 6

The percentage of the variance explained vs the number of principal components: (a) Tunxi dataset, China; (b) Changhua dataset, China.

Figure 6

The percentage of the variance explained vs the number of principal components: (a) Tunxi dataset, China; (b) Changhua dataset, China.

Close modal

Results of different machine-learning models

In this part, we will examine the performance of the different machine-learning models, including individual models, linear regression, ridge regression, Lasso regression, MLP, feature-aware MLP, and the ensemble model.

Experiment details: Here, we present the implementation details, including hyperparameters for all individual models. For the linear regression model, since there is no hyperparameter, we directly use the optimal solution of the ordinary least square. For ridge regression and Lasso regression, we use grid search to tune the coefficients of the and regularizations, respectively, and the final optimal coefficients are and , respectively. For MLP, we use four layers, including two hidden layers. The first hidden layer has 128 neurons and the second layer has 64 neurons. For feature-aware MLP, we use three MLPs: one for processing rainfall features, the other for dealing with flood flow features, and the last MLP for predicting. All of these MLPs have three layers, including one hidden layer with 64 neurons. These methods are trained using Adam optimizer with a learning rate of 0.001 for 150 iterations. The ensemble model is formed by a linear regression. Thus, there are no hyperparameters, and we perform it on the validation dataset.

The benefit of feature engineering

First, we will investigate the benefits of the developed feature engineering method. We will evaluate the prediction performance of the MLP and linear regression models using both raw and extracted features. The results, summarized in Table 3, clearly demonstrate that using the extracted features significantly enhances prediction performance across all evaluation metrics. For instance, in the Tunxi Basin, applying feature engineering improves performance on the raw data by over 100% when using either the linear regression model or the MLP model. Similarly, significant improvements are observed in the Changhua Basin.

Table 3

Model performance when using the raw features and extracted features

BasinModelFeatureRMSEMAER-square
Tunxi Basin Linear regression
Linear regression
MLP
MLP 
Raw features
Extracted features Raw features Extracted features 
6.017
3.205
4.601
2.213 
5.413
2.522
4.171
1.727 
−0.275
0.691
0.361
0.852 
Changhua Basin Linear regression
Linear regression
MLP
MLP 
Raw features
Extracted features Raw features Extracted features 
3.782
3.025
3.513
2.997 
2.964
2.851
2.902
2.737 
0.561
0.613
0.573
0.661 
BasinModelFeatureRMSEMAER-square
Tunxi Basin Linear regression
Linear regression
MLP
MLP 
Raw features
Extracted features Raw features Extracted features 
6.017
3.205
4.601
2.213 
5.413
2.522
4.171
1.727 
−0.275
0.691
0.361
0.852 
Changhua Basin Linear regression
Linear regression
MLP
MLP 
Raw features
Extracted features Raw features Extracted features 
3.782
3.025
3.513
2.997 
2.964
2.851
2.902
2.737 
0.561
0.613
0.573
0.661 

RMSE and MAE are measured in hours. Bold font indicates the best values compared with the others.

Performance comparison between different models

Then, we will compare the performances achieved by different models, where the extracted features are used to generate the predictions. The visualizations of the predictions of all models are displayed in Figure 7, where the x-axis represents the ground truth values, which are integers between 0 and 20, and the y-axis represents the predictions (we select the median of predictions when the corresponding real arrival times are the same). For different models, the closer the prediction curve to the line y = x, the smaller the prediction error. Besides, the prediction errors of all models under different metrics are also summarized in Table 4, where the errors are measured in hours. Some findings can be drawn
  • Under these three metrics, it is evident that performing the model ensemble can significantly enhance the performance of individual models. For instance, in Changhua, the Ensemble Model outperforms others with the lowest RMSE (2.564) and MAE (2.192), and the highest R-square (0.735); in Tunxi, the Ensemble Model also shows the best performance with the lowest RMSE (1.966) and MAE (1.524) and the highest R-square (0.883).

  • MLP models generally perform better than linear models, although they lack interpretability to some extent. This implies that, for the arrival time prediction task, using the black-box MLP may typically lead to better performance, when provided with the extracted features.

  • Using ridge regression and Lasso regression performs similarly to the standard linear regression, implying that overfitting or feature redundancy (after the feature cleaning) may not be significant.

Figure 7

Predictions obtained using different machine-learning models for (a) Tunxi and (b) Changhua dataset. The x-axis represents the real arrival time in hour, the y-axis represents the predicted arrival time in hour. The closer to the line y = x, the smaller the prediction error.

Figure 7

Predictions obtained using different machine-learning models for (a) Tunxi and (b) Changhua dataset. The x-axis represents the real arrival time in hour, the y-axis represents the predicted arrival time in hour. The closer to the line y = x, the smaller the prediction error.

Close modal
Table 4

Performance comparison between different models

BasinModelRMSEMAER-square
Tunxi Basin Linear regression 3.205 2.522 0.691 
Ridge regression 3.348 3.044 0.662 
Lasso regression 3.343 3.038 0.663 
MLP model 2.213 1.727 0.852 
Feature-aware MLP 2.864 2.290 0.752 
Ensemble model 1.966 1.524 0.883 
Changhua Basin Linear regression 3.025 2.851 0.613 
Ridge regression 3.013 2.808 0.622 
Lasso regression 3.005 2.787 0.629 
MLP model 2.997 2.737 0.661 
Feature-aware MLP 3.040 2.891 0.609 
Ensemble model 2.564 2.192 0.735 
BasinModelRMSEMAER-square
Tunxi Basin Linear regression 3.205 2.522 0.691 
Ridge regression 3.348 3.044 0.662 
Lasso regression 3.343 3.038 0.663 
MLP model 2.213 1.727 0.852 
Feature-aware MLP 2.864 2.290 0.752 
Ensemble model 1.966 1.524 0.883 
Changhua Basin Linear regression 3.025 2.851 0.613 
Ridge regression 3.013 2.808 0.622 
Lasso regression 3.005 2.787 0.629 
MLP model 2.997 2.737 0.661 
Feature-aware MLP 3.040 2.891 0.609 
Ensemble model 2.564 2.192 0.735 

RMSE and MAE are measured in hours.

These findings demonstrate the effectiveness of our proposed machine-learning framework and justify the necessity of each module within it.

In this section, we would like to discuss the key findings and a number of limitations of our system. In particular, one key finding is that, according to the analysis of the different features in Table 2, the historical rainfall data are more crucial for predicting the flood peak arrival time compared to the historical flood flow. This could be useful for other flood-based prediction tasks. The other key finding is that model ensemble is crucial in real-world forecasting tasks. If there is a better individual prediction model, we are able to incorporate it into our ensemble framework to further improve the performance.

Regarding the limitations, one of them is that our feature engineering method only considers the historical rainfall and flood flow. This is restricted by the Tunxi and Changhua datasets. Another aspect of our framework that can be improved is the design of machine-learning models. In many real-world scenarios, we are able to collect data from multiple sources/stations at different locations in the Basin, and then it naturally forms a spatial–temporal data structure (Shao et al. 2023). Our current model does not take the spatial correlation of the data into consideration, so there may still be space for further improvement. Last but not least, our method is data driven. The lack of physical understanding and constraints may lead to unexpected and unreasonable predictions in some extreme cases.

In this paper, we studied an important flood forecasting problem: predicting the arrival time of flood peaks, which is essential for timely warnings and preparations but has received limited attention. We further developed a comprehensive machine-learning framework with novel designs for accomplishing the forecasting task. Specifically, the framework is built based on the combination of an interpretable feature engineering method, individual machine-learning models, and a novel model ensemble technique. This paper further applies the proposed machine-learning framework to predict the flood peak arrival time in the Tunxi and Changhua basins, China. Numerical results show the effectiveness of our forecasting system. Feature engineering improves MAE in the Tunxi Basin from 5.413 to 2.522 for the linear model and from 4.171 to 1.727 for the MLP model. Our ensemble methods outperform baselines, achieving MAE errors of 1.524 in Tunxi and 2.192 in Changhua, compared to 1.727 and 2.737 from the best baselines. These advancements enable more effective flood management through proactive measures and timely responses.

We also identify several future directions that can further improve the prediction accuracy. First, it is interesting to discover more informative features if other observations can be obtained, such as temperature, evaporation, soil moisture content, etc. Then, the machine-learning model can also be improved to involve the spatial component, such as graph neural networks-based models, to make better use of the spatial–temporal data. Moreover, in order to achieve more reliable and controllable prediction, a future direction is to incorporate physical understandings into the model design, i.e., integrating the conventional hydrological models of flood flow into the data-driven machine-learning models.

This research was funded by the Research Projects of the Department of Education of Guangdong Province, China grant number 2023ZDZX1083, and the Research Foundation of Shenzhen Polytechnic University, China under grant number 6023312050K.

Data cannot be made publicly available; readers should contact the corresponding author for details.

The authors declare there is no conflict.

Agarwal
S.
,
Roy
P. J.
,
Choudhury
P.
&
Debbarma
N.
(
2021a
)
Flood forecasting and flood flow modeling in a river system using ANN
,
Water Practice and Technology
,
16
(
4
),
1194
1205
.
doi:10.2166/wpt.2021.068
.
Agarwal
S.
,
Roy
P. J.
,
Choudhury
P. S.
&
Debbarma
N.
(
2021b
)
River flow forecasting by comparative analysis of multiple input and multiple output models form using ANN
,
H2Open Journal
,
4
(
1
),
413
428
.
doi:10.2166/h2oj.2021.122
.
Agarwal, S., Roy, P., Choudhury, P. & Debbarma, N. (2022) Comparative study on stream flow prediction using the GMNN and wavelet-based GMNN, Journal of Water and Climate Change, 13 (9), 3323–3337. doi:10.2166/wcc.2022.226.
Agarwal S., Mukherjee, D. & Debbarma, N. (2023) Analysis of extreme annual rainfall in North-Eastern India using machine learning techniques, AQUA – Water Infrastructure, Ecosystems and Society, 72 (12), 2201–2215. doi:10.2166/aqua.2023.016.
Ashok
S. P.
&
Pekkat
S.
(
2022
)
A systematic quantitative review on the performance of some of the recent short-term rainfall forecasting techniques
,
Journal of Water and Climate Change
,
13
(
8
),
3004
3029
.
doi:10.2166/wcc.2022.302
.
Atashi
V.
,
Gorji
H. T.
,
Shahabi
S. M.
,
Kardan
R.
&
Lim
Y. H.
(
2022
)
Water level forecasting using deep learning time-series analysis: a case study of red river of the north
,
Water
,
14
(
12
),
1971
.
doi:10.3390/w14121971
.
Badrzadeh
H.
,
Sarukkalige
R.
&
Jayawardena
A.
(
2015
)
Hourly runoff forecasting for flood risk management: application of various computational intelligence models
,
Journal of Hydrology
,
529
,
1633
1643
.
doi:10.1016/j.jhydrol.2015.07.057
.
Bracher, J., Wolffram, D., Deuschel, J., Görgen, K., Ketterer, J. L., Ullrich, A., Abbott, S., Barbarossa, M. V., Bertsimas, D., Bhatia, S., Bodych, M., Bosse, N. I., Burgard, J. P., Castro, L., Fairchild, G., Fuhrmann, J., Funk, S., Gogolewski, K., Gu, Q., Heyder, S., Hotz, T., Kheifetz, Y., Kirsten, H., Krueger, T., Krymova, E., Li, M. L., Meinke, J. H., Michaud, I. J., Niedzielewski, K., Ożański, T., Rakowski, F., Scholz, M., Soni, S., Srivastava, A., Zieliński, J., Zou, D., Gneiting, T., Schienle, M. & List of Contributors by Team
(
2021
)
A pre-registered short-term forecasting study of COVID-19 in Germany and Poland during the second wave
,
Nature Communications
,
12
(
1
),
5173
.
doi:10.1038/s41467-021-25207-0
.
Carlos
H. R. L.
&
Upmanu
L.
(
2010
)
Climate informed monthly streamflow forecasts for the Brazilian hydropower network using a periodic ridge regression model
,
Journal of Hydrology
,
380
(
3–4
),
438
449
.
doi:10.1016/j.jhydrol.2009.11.016
.
Chakraborty
S.
&
Biswas
S.
(
2023
)
River discharge prediction using wavelet-based artificial neural network and long short-term memory models: a case study of Teesta River Basin, India
,
Stochastic Environmental Research and Risk Assessment
,
37
(
8
),
3163
3184
.
doi:10.1007/s00477-023-02443-y
.
Deng
P.
&
Li
Z.
(
2013
)
Comparison of three hydrological models in flood simulation for Xixian Basin of Huaihe River
,
Journal of Hohai University (Natural Sciences)
,
41
(
5
),
377
382
.
doi:10.3876/j.issn.1000-1980.2013.05.001
.
Ding
Y.
,
Zhu
Y.
,
Feng
J.
,
Zhang
P.
&
Cheng
Z.
(
2020
)
Interpretable spatio-temporal attention LSTM model for flood forecasting
,
Neurocomputing
,
403
,
348
359
.
doi:10.1016/j.neucom.2020.04.110
.
Garg
N.
,
Negi
S.
,
Nagar
R.
,
Rao
S.
&
S
K. R.
(
2023
)
Multivariate multi-step LSTM model for flood runoff prediction: a case study on the Godavari River Basin in India
,
Journal of Water and Climate Change
,
14
(
10
),
3635
3647
.
doi:10.2166/wcc.2023.374
.
Gary
C. M.
(
2009
)
Ridge regression
,
Wiley Interdisciplinary Reviews: Computational Statistics
,
1
(
1
),
93
100
.
doi:10.1002/wics.14
.
Han
S.
&
Coulibaly
P.
(
2017
)
Bayesian flood forecasting methods: a review
,
Journal of Hydrology
,
551
,
340
351
.
doi:10.1016/j.jhydrol.2017.06.004
.
Hapuarachchi
H.
,
Wang
Q.
&
Pagano
T.
(
2011
)
A review of advances in flash flood forecasting
,
Hydrological Processes
,
25
(
18
),
2771
2784
.
doi:10.1002/hyp.8040
.
Haribabu
S.
,
Gupta
G. S.
,
Kumar
P. N.
&
Rajendran
P. S.
(
2021
). ‘
Prediction of flood by rainfall using MLP classifier of neural network model
',
2021 6th International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India
, pp.
1360
1365
.
doi: 10.1109/ICCES51350.2021.9489161
.
Hayder
I. M.
,
Al-Amiedy
T. A.
,
Ghaban
W.
,
Saeed
F.
,
Nasser
M.
,
Al-Ali
G. A.
&
Younis
H. A.
(
2023
)
An intelligent early flood forecasting and prediction leveraging machine and deep learning algorithms with advanced alert system
,
Processes
,
11
(
2
),
481
.
doi:10.3390/pr11020481
.
Hervé
A.
&
Williams
L. J.
(
2010
)
Principal component analysis
,
Wiley Interdisciplinary Reviews: Computational Statistics
,
2
(
4
),
433
459
.
doi:10.1002/wics.101
.
James
G.
,
Witten
D.
,
Hastie
T.
,
Tibshirani
R.
&
Taylor
J.
(
2023
)
Linear Regression
. In: James, G., Witten, D., Hastie, T., Tibshirani, R. & Taylor, J. (eds)
An Introduction to Statistical Learning (Springer Texts in Statistics)
,
Cham
:
Springer
, pp.
61
89
.
doi:10.1007/978-3-031-38747-0_3
.
Jia
S.
,
Li
Y.
,
A.
,
Liu
W.
,
Zhu
W.
,
Yan
J.
,
Liang
Y.
,
Xiang
X.
&
Guan
Z.
(
2018
)
City storm-flood events in China, 1984–2015
,
International Journal of Water Resources Development
,
35
(
4
),
605
618
.
doi:10.1080/07900627.2018.1513830
.
Kabbilawsh
P.
,
Kumar
D. S.
&
Chithra
N. R.
(
2022
)
Performance evaluation of univariate time-series techniques for forecasting monthly rainfall data
,
Journal of Water and Climate Change
,
13
(
12
),
4151
4176
.
doi:10.2166/wcc.2022.107
.
Kim
M.
,
Lee
S.
&
Jeong
T.
(
2023
)
Time series prediction methodology and ensemble model using real-world data
,
Electronics
,
12
(
13
),
2811
.
doi:10.3390/electronics12132811
.
Le
X. H.
,
Ho
H. V.
,
Lee
G.
&
Jung
S.
(
2019
)
Application of long short-term memory (LSTM) neural network for flood forecasting
,
Water
,
11
,
1387
.
doi:10.3390/w11071387
.
Lee
J. Y.
&
Kim
J. S.
(
2021
)
Detecting areas vulnerable to flooding using hydrological-topographic factors and logistic regression
,
Applied Sciences
,
11
(
12
),
5652
.
doi:10.3390/app11125652
.
Li
Y.
,
Gong
S.
,
Zhang
Z.
,
Liu
M.
,
Sun
C.
&
Zhao
Y.
(
2021
)
Vulnerability evaluation of rainstorm disaster based on ESA conceptual framework: a case study of Liaoning Province, China
,
Sustainable Cities and Society
,
64
,
102540
.
doi:10.1016/j.scs.2020.102540
.
Meteorological Service of Canada
(
2013
)
MANOBS-Manual of Surface Weather Observations
, 7th edn.
Amendment
, p.
18
. .
Ming
X.
,
Liang
Q.
,
Xia
X.
,
Li
D.
&
Fowler
H. J.
(
2020
)
Real-time flood forecasting based on a high-performance 2-D hydrodynamic model and numerical weather predictions
,
Water Resources Research
,
56
(
7
),
e2019WR025583
.
doi:10.1029/2019WR025583
.
Mohammad
R. P.-R.
,
Dahmardeh
K.
,
Hadizadeh
M.
,
Keykha
G.
,
Mohammadnia
N.
,
Gangali
M.
,
Keikha
M.
,
Davatgar
N.
&
Brungard
C.
(
2020
)
Prediction of soil water infiltration using multiple linear regression and random forest in a dry flood plain, Eastern Iran
,
CATENA
,
194
,
104715
.
doi:10.1016/j.catena.2020.104715
.
Mumtaz
A.
,
Prasad
R.
,
Xiang
Y.
&
Yaseen
Z. M.
(
2020
)
Complete ensemble empirical mode decomposition hybridized with random forest and kernel ridge regression model for monthly rainfall forecasts
,
Journal of Hydrology
,
584
,
124647
.
doi:10.1016/j.jhydrol.2020.124647
.
Nguyen
D. T.
&
Chen
S. T.
(
2020
)
Real-time probabilistic flood forecasting using multiple machine learning methods
,
Water
,
12
(
3
),
787
.
doi:10.3390/w12030787
.
Noor
F.
,
Haq
S.
,
Rakib
M.
,
Ahmed
T.
,
Jamal
Z.
,
Siam
Z. S.
,
Hasan
R. T.
,
Adnan
M. S. G.
,
Dewan
A.
&
Rahman
R. M.
(
2022
)
Water level forecasting using spatiotemporal attention-based long short-term memory network
,
Water
,
14
(
4
),
612
.
doi:10.3390/w14040612
.
Pablo
M.
&
Dormido-Canto
S.
(
2023
)
A novel feature engineering approach for high-frequency financial data
,
Engineering Applications of Artificial Intelligence
,
125
,
106705
.
doi:10.1016/j.engappai.2023.106705
.
Palash
W.
,
Jiang
Y.
,
Akanda
A. S.
,
Small
D. L.
,
Nozari
A.
&
Islam
S.
(
2018
)
A streamflow and waterlevel forecasting model for the Ganges, Brahmaputra, and Meghna Rivers with requisite simplicity
,
Journal of Hydrometeorology
,
19
(
1
),
201
225
.
doi:10.1175/JHM-D-16-0202.1
.
Pedro
H. B.
,
Zakordonets
O.
&
Teixeira
J. P.
(
2021
)
A COVID-19 time series forecasting model based on MLP ANN
,
Procedia Computer Science
,
181
,
940
947
.
doi:10.1016/j.procs.2021.01.250
.
Rahman
T.
,
Syeed
M. M. A.
,
Farzana
M.
,
Namir
I.
,
Ishrar
I.
,
Nushra
M. H.
&
Khan
B. M.
(
2023
) ‘
Flood prediction using ensemble machine learning model
',
Proceedings of the 2023 5th International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA), Istanbul, Turkey
, pp.
1
6
.
doi:10.1109/HORA58378.2023.10156673
.
Rajab
A.
,
Farman
H.
,
Islam
N.
,
Syed
D.
,
Elmagzoub
M.
,
Shaikh
A.
,
Akram
M.
&
Alrizq
M.
(
2023
)
Flood forecasting by using machine learning: a study leveraging historic climatic records of Bangladesh
,
Water
,
15
(
22
),
3970
.
doi:10.3390/w15223970
.
Ranstam
J.
&
Cook
J. A.
(
2018
)
LASSO regression
,
British Journal of Surgery
,
105
(
10
),
1348
.
doi:10.1002/bjs.10895
.
Rastogi
A.
,
Qais
A.
,
Saxena
A.
&
Sinha
D.
(
2021
) ‘
Stock market prediction with Lasso regression using technical analysis and time lag
',
2021 6th International Conference for Convergence in Technology (I2CT), Maharashtra, India
, pp.
1
5
.
doi:10.1109/I2CT51068.2021.9417935
.
Rofah
N. R.
,
Sari
A. C.
&
Yohanes
(
2021
)
Lasso regression for daily rainfall modeling at Citeko Station, Bogor, Indonesia
,
Procedia Computer Science
,
179
,
383
390
.
doi:10.1016/j.procs.2021.01.020
.
Santos
C. A. G.
,
do Nascimento
G. R.
,
de Farias
C. A. S.
,
da Silva
R. M.
&
Mishra
M.
(
2023
)
Short- and long-term streamflow forecasting using wavelet neural networks for complex watersheds: a case study in the Mahanadi River, India
,
Ecological Informatics
,
73
,
101945
.
doi: 10.1016/j.ecoinf.2022.101945
.
Shao
P.
,
Feng
J.
,
Wu
Y.
,
Wang
W.
&
Lu
J.
(
2023
)
Data-driven and knowledge-guided heterogeneous graphs and temporal convolution networks for flood forecasting
,
Applied Sciences
,
13
(
12
),
7191
.
doi:10.3390/app13127191
.
Sharma
S.
&
Kumari
S.
(
2024
)
Comparison of machine learning models for flood forecasting in the Mahanadi River Basin, India
,
Journal of Water and Climate Change
,
15
(
4
),
1629
1652
.
doi:10.2166/wcc.2024.517
.
Shijun
C.
,
Qin
W.
,
Yanmei
Z.
,
Guangwen
M.
,
Xiaoyan
H.
&
Liang
W.
(
2020
)
Medium-and long-term runoff forecasting based on a random forest regression model
,
Water Supply
,
20
(
8
),
3658
3664
.
doi:10.2166/ws.2020.214
.
Suddul
G.
,
Dookhitram
K.
,
Bekaroo
G.
&
Shankhur
N.
(
2020
) ‘
An evolutionary multilayer perceptron algorithm for real time river flood prediction
',
2020 Zooming Innovation in Consumer Technologies Conference (ZINC), Novi Sad, Serbia
, pp.
109
112
.
doi:10.1109/ZINC50678.2020.9161824
.
Taud
H.
,
Mas
J.
, (
2018
)
Multilayer Perceptron (MLP)
. In:
Camacho Olmedo
M.
,
Paegelow
M.
,
Mas
J. F.
&
Escobar
F.
(eds.)
Geomatic Approaches for Modeling Land Change Scenarios, Lecture Notes in Geoinformation and Cartography
,
Cham
:
Springer
.
doi:10.1007/978-3-319-60801-3_27
.
Toth
E.
,
Brath
A.
&
Montanari
A.
(
2000
)
Comparison of short-term rainfall prediction models for real-time flood forecasting
,
Journal of Hydrology
,
239
(
1–4
),
132
147
.
doi:10.1016/S0022-1694(00)00344-9
.
Tsakiri
K.
,
Marsellos
A.
&
Kapetanakis
S.
(
2018
)
Artificial neural network and multiple linear regression for flood prediction in Mohawk River, New York
,
Water
,
10
,
1158
.
doi:10.3390/w10091158
.
Wang
C.
,
Baratchi
M.
,
Bäck
T.
,
Hoos
H. H.
,
Limmer
S.
&
Olhofer
M.
(
2022
)
Towards time-series feature engineering in automated machine learning for multi-step-ahead forecasting
,
Engineering Proceedings
,
18
(
1
),
17
.
doi:10.3390/engproc2022018017
.
Xu
J.
,
Li
Z.
,
Huo
W.
&
Ma
Y.
(
2019
)
Comparison of real-time correction methods of flood forecasting in semi-humid watershed
,
Journal of Hohai University (Natural Sciences)
,
47
(
4
),
317
322
.
doi:10.3876/j.issn.1000-1980.2019.04.005
.
Yenming, C., Jianquan, Z. & Yan, M. (2016) Flood forecasting by ensemble neural networks, Journal of Zhejiang University (Engineering Science), 50 (8), 1471–1478. doi:10.3785/j.issn.1008-973X.2016.08.007.
Zhao
N.
,
Wang
H.
&
Zhang
B.
(
2019
)
Runoff variation in Zoige wetland basin and its response to climate change
,
Water Resources Protection
,
35
(
5
),
40
47
.
doi:10.3880/j.issn.1004-6933.2019.05.008
.
Zhihui
D.
,
Ming
Z.
,
Nadia
N.
,
Dong
X.
&
Feng
Y.
(
2023
)
A hydrological data prediction model based on LSTM with attention mechanism
,
Water
,
15
(
4
),
670
.
doi:10.3390/w15040670
.
This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (CC BY 4.0), which permits copying, adaptation and redistribution, provided the original work is properly cited (http://creativecommons.org/licenses/by/4.0/).