The forecasting of reference crop evapotranspiration (ET0) plays a crucial role in irrigation scheduling. In this study, we employed deep learning (DL) models, including convolutional neural network (CNN), long short-term memory (LSTM), bidirectional LSTM (Bi-LSTM), CNN-LSTM, and CNN-Bi-LSTM, to forecast ET0 at three different stations (Nanchang, Xinjian, and Dongxiang) in Jiangxi Province, China, for various forecast lead times (1, 3, 5, 7, 10, and 15 days). The experimental results demonstrate that LSTM and Bi-LSTM achieved the highest predictive accuracy across the three stations, with corresponding average values of coefficient of determination (R²), mean absolute error (MAE), root mean square error (RMSE), Nash–Sutcliffe efficiency (NSE), and combined accuracy (CA) of 0.9646, 0.2456 mm day−1, 0.2931 mm day−1, 0.9646, and 0.2043, respectively. Compared with the baseline model, the hybrid models incorporating complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN), EEMD, and CEEMD exhibited inferior predictive performance. Additionally, their significantly longer runtime resulted in operational costs that were 8–10 times higher. In contrast, DL-based LSTM and Bi-LSTM are excellent algorithms for forecasting ET0 across various time intervals.

  • Accurately predicted the evapotranspiration under different cycles by deep learning LSTM and Bi-LSTM models based on the time-series structure.

  • Revealed the potential of combining ensemble empirical mode decomposition, complete ensemble empirical mode decomposition, and complete ensemble empirical mode decomposition with adaptive noise with deep learning for evapotranspiration prediction.

Evapotranspiration (ET) is the physical process by which liquid water is transferred from the soil and plant surfaces into the atmosphere as water vapor. It is a crucial factor in the interactions among soil, plants, and the atmosphere, playing a significant role in the water balance and energy equilibrium at the Earth's surface (Allen et al. 1998; Liu et al. 2013). Moreover, is an indispensable component in the terrestrial water, energy, and carbon cycles, making it crucial for water resource planning and management (Tao et al. 2018). To make irrigation systems work better, accurate predictions of crop evapotranspiration are needed. These predictions need to cover a long enough time in the future to take into account changes in water supply over time (Torres et al. 2011a). Similarly, accurate forecasting of crop water requirements (i.e., actual evapotranspiration, ) is crucial for designing and planning irrigation scheduling (Perera et al. 2014). For large-scale irrigation system forecasting, Cetin et al. (2023) used the mapping evapotranspiration at high resolution and the internalized calibration (METRIC) model to estimate using satellite and meteorological data, highlighting the importance of estimating for enhancing water efficiency.

Crop evapotranspiration () can be estimated by multiplying the reference evapotranspiration () with the crop coefficient . represents the ratio between (Jensen 1968). Alternatively, can be measured directly using an atmometer. However, this method is time-consuming and costly, making it less suitable for research purposes (Wright 1988; Tao et al. 2018). Simultaneously, research on data is also highly noteworthy. Başakın et al. systematically summarized the process of estimating missing data. From these studies, the complexity and nonlinearity of data can be inferred (Başakın et al. 2023). Currently, the most well-known mathematical formula for determining is the FAO-56 Penman–Monteith (PM) equation. This formula was proposed by the Food and Agriculture Organization of the United Nations (FAO) (Allen et al. 1998). The FAO-56 PM model, taking into account both thermodynamic and aerodynamic effects, is more accurate than existing empirical models. As a result, the FAO-56 PM method has been widely applied as a reference method in various regions and climates (Tao et al. 2018; Zhang et al. 2018; Granata 2019; Tikhamarine et al. 2019; Wu et al. 2019b; Lu et al. 2023). However, this model has a notable drawback. It requires a substantial amount of input meteorological data to be effectively utilized, including parameters such as maximum/minimum temperature, wind speed, relative humidity (RH), and solar radiation (Feng et al. 2017b). Therefore, the FAO-56 PM model can only be optimally employed when comprehensive meteorological data is available. For scenarios with limited meteorological data, it is advisable to use models that require fewer meteorological variables (Wen et al. 2015). In the past few decades, efforts have been made to address the issue of requiring extensive meteorological inputs by relying on empirical models (Mehdizadeh et al. 2017). Pendiuk et al. reduced the data requirements for estimating by utilizing superconducting gravimeter data to calculate cumulative (Pendiuk et al. 2023). In addition, many studies based on radiation, humidity, and temperature, and have aimed to estimate reference evapotranspiration using limited meteorological data (Xu & Singh 2002; Bickici Arikan et al. 2021). For example, there are temperature-based methods such as the Hargreaves and the modified Hargreaves methods (Hargreaves & Samani 1985; Luo et al. 2014), radiation-based methods including the Priestley–Taylor, Marking, and Ritchie models (Priestley & Taylor 1972; Feng et al. 2016), as well as the mass-transfer-based Trabert model (Trabert 1896). Additionally, these models are more suitable for forecasting on weekly and monthly cycles and may not be as well-suited for daily forecasting (Torres et al. 2011a). We employ Chen's nonlinear theoretical model within a dual-source trapezoid framework based on dry/wet edges to estimate land surface ET. This model enhances the estimation of daily-scale , thereby improving the prediction accuracy of regional evapotranspiration and its components. This research is closer to a physical model for prediction (Chen et al. 2023). These models involve complex physical processes, and predictive accuracy still has room for improvement.

In contrast to attempts with empirical models, researchers have sought to address the challenges associated with empirical models by employing data-driven artificial intelligence (AI) and machine learning (ML) methods to establish mathematical models for forecasting . Due to the high nonlinearity of meteorological data, traditional empirical models struggle to accurately capture the relationships between meteorological variables, thereby hindering their ability to forecast effectively. However, ML does not require a clear understanding of data variables in advance and can offer simple solutions to nonlinear and multivariate problems (Kisi 2015; Wang et al. 2017). Due to its capability to operate without prior knowledge of the underlying data and physical processes and its ability to handle nonlinearity, ML becomes robust, efficient, and reliable in problem-solving and decision-making (Goyal et al. 2023). In past research, various ML techniques have been applied to forecast . For instance, Ferreira et al. used artificial neural networks (ANNs) and support vector machines (SVMs) with limited meteorological data to forecast at 203 stations in Brazil. The results indicated that ANN and SVM exhibited good forecasting performance (Ferreira et al. 2019). Vosoughifar et al. established equations for coastal areas using multi-adaptive regression splines (MARS) and genetic expression programming (GEP), discovering they possess a more accurate estimation capability than traditional regression-based formulas (Vosoughifar et al. 2023). In addition to that, other models such as M5 model trees (Pal & Deswal 2009; Rahimikhoob 2014; Kisi & Kilic 2016), random forest (RF) (Feng et al. 2017a), gradient boosting decision tree (GBDT) (Fan et al. 2018; Wu et al. 2019a), and ANN-based models like multi-layer perceptron (MLP) (Torres et al. 2011a; Ladlani et al. 2014; Traore et al. 2016), generalized regression neural network (KiŞI 2006), radial basis function neural network (RBFNN) (Trajkovic 2005; Ladlani et al. 2012), and extreme learning machine (ELM) (Abdullah et al. 2015; Gocic et al. 2016) have been employed to forecast . In previous studies, to further improve the forecasting accuracy of within the framework of ML, bio-inspired optimization algorithms have been incorporated. For example, in the forecasting of for arid and humid regions in China, a hybrid model combining the whale optimization algorithm (WOA) with extreme gradient boosting (XGBoost) was employed. The results indicated that this strategy could provide a more accurate estimation of daily evaporation (Yan et al. 2021). Wu et al. utilized four bio-inspired optimization algorithms – ant colony optimization, genetic algorithm, flower pollination algorithm, and cuckoo search algorithm – to optimize the ELM model for daily forecasting. The results indicated that the ELM-FPA and ELM-CAS models outperformed the base ELM model regarding forecasting accuracy (Wu et al. 2019b). Zhao et al. (2022) combined two bio-inspired optimization algorithms, golden eagle optimization (GEO) and sparrow search algorithm (SSA), with ELM. The ELM model with the SSA was considered the optimal model for forecasting .

While conventional ML models have made significant strides in the daily forecasting of , there are still limitations to these theoretical models when it comes to forecasting with higher nonlinearity, larger datasets, and more complex cycles. In recent years, deep learning (DL) has been applied to various fields, and its performance has surpassed that of traditional ML models (Shi et al. 2018; Ghimire et al. 2019; Du et al. 2021; Luo et al. 2021; Yan et al. 2023). Furthermore, a growing body of research employs DL for the time-series forecasting of . This differs from traditional forecasting methods, as time-series forecasting involves forecasting a sequence using a sequence, distinct from the conventional regression task where independent variables are used to forecast dependent variables. Time-series forecasting also provides a better understanding of the forecasted trends and aligns well with real-world variations. Sabanci et al. (2023) predicted for 12 stations with variable climatic characteristics in the Central Anatolia region (CAR) using long short-term memory (LSTM), ANN, and MARS models. The results indicated that LSTM, ANN, and ML-based MARS showed high accuracy in estimating . Yin et al. (2020) forecasted daily based on a limited set of meteorological variables using a hybrid bidirectional LSTM (Bi-LSTM) model. The results indicated that the Bi-LSTM model yielded the best forecasting for all stations. Saggi & Jain (2019) estimated reference evapotranspiration in the northern region of Punjab, India, using a DL-MLP in comparison with traditional ML models such as generalized linear model (GLM), RF, and gradient boosting machine (GBM). The results indicated that the DL model achieved the best forecasting performance. Additionally, by combining DL models with convolutional neural networks (CNNs), one can further enhance the training of DL models through the convolutional calculations of CNN. For instance, Ferreira & da Cunha (2020a) used a hybrid CNN-LSTM model to forecast lagged , combining a one-dimensional CNN with an LSTM model. The results suggest that CNN-LSTM performs slightly better than RF and ANN but does not demonstrate the highest accuracy. In addition, combining ML with metaheuristic algorithms to optimize model parameters is also a common strategy. For example, Yuan et al. optimized the parameters of an LSTM model using the ant lion optimization (ALO) algorithm (LSTM-ALO) for monthly runoff prediction (Yuan et al. 2018). Ikram et al. utilized the reptile search algorithm and the weighted vector optimizer to enhance the performance of the LSTM model (LSTM-INFO) in predicting water temperature (Ikram et al. 2023). The application of DL in forecasting still requires further analysis and exploration to maximize its inherent value.

Despite extensive studies and proven effectiveness of ML and DL models in forecasting , bias toward data dependency persists. If 's increased nonlinearity, uncertainty, and randomness fail to capture relevant features/patterns in the time series, including trends, periodicity, seasonality, anomalies, and abrupt change peaks, it becomes increasingly difficult for DL to produce accurate forecasts (Ali et al. 2023). By decomposing the original data to obtain signal components and forecasting each signal component separately, even though there are multiple features in actual problems, each feature undergoes one decomposition at a time (Prasad et al. 2019). The decomposed signals become smoother with more robust periodicity, making them better suited for DL forecasting. Ali et al. employed the multivariate variational mode decomposition (MVMD) and monotonic empirical mode decomposition (MEMD) to decompose the original data. They then used a combination of RF, convolutional feedforward neural network (CFNN), ELM, and boosted regression trees (BRT) to forecast daily . The results indicated that MVMD-BRT, compared with the baseline model (BRT), could provide more accurate daily forecasting accuracy (Ali et al. 2023). Zheng et al. employed the MVMD method combined with the soft feature filter (SoFeFilter) and gated recurrent unit (GRU) model to predict for t + 1 day. Experimental results demonstrated that at the Gympie station, the MVMD-SoFeFilter-GRU model achieved the highest performance metrics. These findings confirm that the MVMD-SoFeFilter-GRU model provides the most accurate predictions for 1 day ahead (Zheng et al. 2023). Similarly, by integrating LSTM with the MEMD method and the Boruta-RF algorithm, ET was forecasted for drought-prone areas in Queensland, Australia. The results indicated that the MEMD-Boruta-LSTM hybrid model outperformed standalone models such as deep neural network (DNN), decision tree (DT), and LSTM, achieving the highest predictive performance across all study locations (Jayasinghe et al. 2021). In addition, time-varying correlation decomposition methods have been applied to forecasting. Karbasi et al. employed the time-varying filter-based empirical mode decomposition (TVF-EMD) technique, coupled with the partial autocorrelation function (PACF), to calculate significant lag values from the decomposed subsequences (i.e., IMFs). The extra tree-Boruta feature selection algorithm was then used to extract key features from the IMFs. By integrating TVF-EMD with the bidirectional recurrent neural network (Bi-RNN), MLP, RF, and XGBoost, weekly evaporation was predicted and evaluated using various performance metrics. The results demonstrated that, compared to models without decomposition, the TVF-EMD hybrid ML models achieved superior forecasting accuracy for weekly at the Redcliffe and Gold Coast stations (Karbasi et al. 2023). Lu et al. combined the variational mode decomposition (VMD) algorithm with the backpropagation neural network (BPNN) model. The results showed that the VMD-BPNN hybrid forecasting model outperformed the empirical mode decomposition (EMD) algorithm with the BPNN (EMD-BPNN) and the ensemble empirical mode decomposition (EEMD) algorithm with BPNN (EEMD-BPNN) models. Additionally, the forecasting accuracy of the VMD-BPNN model was significantly higher than that of single models such as BPNN, support vector regression (SVR), and gradient boosting regression tree (GBRT) (Lu et al. 2023). Commonly used decomposition algorithms include Fourier spectrum analysis (Soman et al. 2015) and discrete wavelet transform (Mallat 1989). Karbasi et al. proposed a time-series decomposition technique known as empirical Fourier decomposition (EFD). Autocorrelation analysis was employed to determine significant lag values, and EFD was applied to decompose the data. Based on the EFD results, lagged data were created, and the K-best feature selection algorithm was used to identify key features. Models were evaluated using the correlation coefficient (R) and root mean square error (RMSE). The results demonstrated that incorporating EFD-decomposed data significantly enhanced the predictive performance of ML models (Karbasi et al. 2024). Katipoğlu estimated monthly ET values for the Hakkari Province by combining SVM, bagging trees, and boosting trees with wavelet transformation. Model performance was evaluated using RMSE, MAE, the coefficient of determination (R2), and Taylor diagrams. The results indicated that hybrid wavelet-ML models, constructed by decomposing input combinations into subcomponents via wavelet transformation, generally produced more accurate predictions compared with standalone ML models (Katipoğlu 2023). Additionally, a method different from the decompositions mentioned above is complete ensemble empirical mode decomposition adaptive noise (CEEMDAN) (Yeh et al. 2010), which has been applied to many other fields with satisfactory results (Zhang et al. 2017; Gao et al. 2020; Yan et al. 2023). It is generally believed that decomposition methods have advantages, but in geographically complex environments, where numerous influencing factors and frequent meteorological changes occur, time series-based methods can be risky. This study verifies this risk from multiple perspectives. However, based on the author's understanding, there is currently no specific analysis of the application of CEEMDAN combined with CNN and DL for forecasting . A detailed analysis of this algorithm is of significant reference value for future research on forecasting.

Due to the nonlinearity among climate variables, accurately predicting has become a challenge. Related literature shows few studies on DL-based models combined with CEEMDAN models to forecast in China. This study proposes four new methods for forecasting using CNN, LSTM, Bi-LSTM, and CNN-Bi-LSTM combined with CEEMDAN. This study compared hybrid models that integrate different decomposition algorithms, including EEMD, CEEMD, and CEEMDAN, with standalone model approaches, offering a more accurate forecasting strategy for multi-period prediction. Therefore, the objectives of this study are as follows: (1) Determine the importance and forecasting potential of CEEMDAN combined with base models XGBoost, SVM, RF, CNN, LSTM, Bi-LSTM, CNN-LSTM, and CNN-Bi-LSTM for forecasting; (2) Develop the base model and CEEMDAN models for forecasting at different periods; (3) Compare the accuracy, stability, and computational cost of baseline models with models based on EEMD, CEEMD, and CEEMDAN for predicting over different periods, and identify the optimal model; (4) Analyze the compatibility of the base model with the combination of CEEMDAN and CNN models for predicting at different periods, aiming to explore potential synergies and optimize overall predictive performance.

Study area and data collection

The meteorological data for this study were collected from three different weather stations in Jiangxi Province, China (Figure 1). Three weather stations were selected to assess the model's generalizability for forecasting in this province: Nanchang, Xinjian, and Dongxiang in Jiangxi Province (Table 1). The research data used in this study were derived from daily maximum temperature (), daily minimum temperature (), sunshine duration (SSD), wind speed (Wind Speed), daily average RH, and annual precipitation (P) from the China Meteorological Administration website (www.cma.com) for the period from 1 September 2001 to 31 August 2021. Table 1 describes the average values of each input meteorological variable and the target variable for the three stations. It also provides specific information about the geographical location of each station, including latitude, longitude, station ID, and elevation. This fundamental analysis is a reference for understanding the meteorological data at the three stations. Additionally, before using the data as input for the model, we performed min–max normalization, constraining the values to be within the range of 0–1. This normalization enhances the model's performance. In this experiment, the dataset was split into a training set and a test set in an 8:2 ratio, and predictions were made on the test set.
Table 1

The geographical location information and the average values of meteorological variables for the three study stations in this research

Station nameStation IDLongitude (E)Latitude (N)Altitude (m) (°C) (°C )SSD (h)Wind ()RH (%)P () ()
Nanchang 58606 115.58 28.40 46.9 22.77 15.85 5.07 1.81 72.72 1,627.6 3.00 
Xinjian 58693 115.53 28.44 40 22.97 16.23 4.64 1.35 70.08 1,546.4 2.86 
Dongxiang 58618 116.35 28.14 50.6 23.19 15.32 4.45 2.36 77.78 1,823.3 2.89 
Station nameStation IDLongitude (E)Latitude (N)Altitude (m) (°C) (°C )SSD (h)Wind ()RH (%)P () ()
Nanchang 58606 115.58 28.40 46.9 22.77 15.85 5.07 1.81 72.72 1,627.6 3.00 
Xinjian 58693 115.53 28.44 40 22.97 16.23 4.64 1.35 70.08 1,546.4 2.86 
Dongxiang 58618 116.35 28.14 50.6 23.19 15.32 4.45 2.36 77.78 1,823.3 2.89 
Figure 1

Map of the three study stations in this paper.

Figure 1

Map of the three study stations in this paper.

Close modal

FAO-56 PM model

The FAO-56 PM equation has been widely used as a reference equation for estimating in various scientific fields, such as agriculture and hydrology (Fan et al. 2018; Chu et al. 2019; Jerin et al. 2021; Zhang et al. 2023). In this study, the daily calculated by the PM equation is used as the target value for evaluating the performance of each model. The mathematical expression of the PM equation is as follows:
(1)
where represents reference evapotranspiration (); is the slope of the vapor pressure curve (kPa °C−1); is the net solar radiation at the Earth's surface (); G is the soil heat flux (); is the psychrometric constant (kPa °C−1); T is the average air temperature at a height of 2 m (°C); is the wind speed at a height of 2 m (); and are the saturation vapor pressure () and the actual vapor pressure (), respectively.

Conventional ML models for forecasting ET0

Researchers have widely applied conventional ML models in forecasting, yielding excellent forecasting results (Fan et al. 2018; Wu et al. 2019a; Ferreira & da Cunha 2020b; Yan et al. 2021). In previous studies, XGBoost (Fan et al. 2018), SVM (Granata 2019), and RF (Ali et al. 2023) have demonstrated stable performance in forecasting. This study used XGBoost, SVM, and RF as baseline models to compare with DL models, thereby enhancing the rigor of the DL forecasting results.

Extreme gradient boosting

XGBoost is an algorithm or engineering implementation based on GBDT proposed by Chen & Guestrin (2016). The basic idea of XGBoost is similar to GBDT, but XGBoost introduces the concept of ‘boosting’, aggregating forecasting from ‘weak’ learners to generate a strong learner. Additionally, XGBoost incorporates second-order derivative optimization, making the loss function more precise. The regularization term optimization helps prevent the overfitting of trees. In addition, XGBoost is an efficient distributed implementation, making its training speed faster.

Support vector machine

SVM is a supervised ML algorithm proposed by Cortes & Vapnik (1995). Based on the principle of structural risk minimization (SRM), SVMs seek to minimize the upper bound of generalization error rather than the upper bound of empirical error. To make predictions in problems that are not linearly related, SVM models can use a set of high-dimensional linear functions to make regression functions (Pai & Lin 2005). SVM employs different kernel functions as a key component to address nonlinear problems. Kernel functions can transform the original features of the data from a lower-dimensional space to a higher-dimensional feature space, making the data linearly separable in the new space. This study utilized the RBF kernel function because, compared with other kernel functions, it demonstrated better performance in forecasting (Kisi 2015).

Random forest

RF was introduced by Breiman (2001). An ensemble learning method utilizes the ‘bagging’ concept to integrate a collection of decision trees with controlled variance. By combining multiple decision trees, RF aims to enhance the accuracy of the model. The ensemble nature of RF, which uses the voting results of multiple decision trees, makes it more robust than a single DT and better able to handle overfitting problems. RF randomly selects the best factor from a subset of forecasters at each node to split it, thereby improving forecasting performance (Liaw & Wiener 2002).

DL models for forecasting ET0

This paper employs time series models, including CNN, LSTM, and Bi-LSTM, which are capable of capturing complex linear relationships in time series data. In this study, multi-period forecasting was conducted, where LSTM better captures long-term temporal dependencies.

Convolutional neural network

CNN has been applied to various fields to address problems, yielding favorable results (Cao et al. 2018; Hoseinzade & Haratizadeh 2019; Viton et al. 2020). Traditional CNN structures consist of three different types of essential layers: convolutional layers, pooling layers, and fully connected layers. For feature extraction, CNN can automatically extract features from input data, which is very different from manually extracting features, and it eliminates the need for data preprocessing (Ferreira & da Cunha 2020b). Moreover, CNN has the characteristic of parameter sharing, where weight parameters are shared at different locations, reducing model complexity, training time, and memory requirements and enhancing robustness. Since CNN is commonly used for image processing problems, it employs 2D convolutional filtering, as images are typically composed of two-dimensional arrays (matrices) (Ferreira & da Cunha 2020a). However, when dealing with time series-based problems in this experiment, CNN with one-dimensional (1D) convolutional filters was used.

Long short-term memory

LSTM is an RNN architecture proposed by Hochreiter & Schmidhuber (1997). DL models, including LSTM, have been shown to outperform shallow ML models and traditional data analysis methods (Janiesch et al. 2021). LSTM addresses the vanishing and exploding gradient problems encountered during backpropagation in time-series forecasting tasks, which were issues associated with traditional RNNs. LSTM introduces a gating mechanism, and its basic unit consists of a memory cell and three adaptive multiplicative gates (input gate, update gate, and forget gate). LSTM primarily operates by multiplying the cell's input with the input gate, selectively forgetting information in the cell state through the forget gate, and ultimately obtaining the next state value by multiplying the activation with the output gate. The mathematical expression for the LSTM model is as follows:
(2)
(3)
(4)
(5)
(6)
(7)
where is the activation function for the LSTM forget gate, input gate, and output gate, with a range from 0 to 1. Tanh is the activation function for the update unit, ranging from −1 to 1. The activation functions control the output of each link. , , , and represent the weight matrices for the forget gate, input gate, update unit, and output gate, respectively. , , , and represent the bias terms in each of the function expressions. represents the output from the output gate to the next memory cell.

Bidirectional long short-term memory

Bi-LSTM is a Bi-RNN introduced by Schuster & Paliwal (1997). Bi-RNN can be trained to forecast only within a predefined future, where each training sequence has both forward and backward RNNs. In such continuous time series tasks, Bi-LSTM can simultaneously capture information from the past and the future. Unlike LSTM, the hidden state in LSTM cannot capture information from the future. The basic structure of Bi-LSTM consists of two LSTMs, where one LSTM is responsible for training on sequentially collected input data from the front, and the other is responsible for training on data collected from the back. This architecture enhances the model's accuracy and performance in handling time series tasks. The hidden layer and the final input of Bi-LSTM can be represented as:
(8)
(9)
(10)
where LSTM is the standard model, is the final output, and represent the forward and backward output weights of the hidden layer, and is the bias term of the hidden layer at time t.

CNN-LSTM and CNN-Bi-LSTM

CNN-LSTM and CNN-Bi-LSTM combine the advantages of CNN with LSTM and Bi-LSTM, making the models more stable. The process adopted in this study for CNN-LSTM and CNN-Bi-LSTM involves taking the previous time series as input data and maintaining consistency in input dimensions with 1D-CNN. The input, convolution, and pooling layers are retained to extract features. The extracted features are then used as input for LSTM and Bi-LSTM, and the final result is forecasted through LSTM and Bi-LSTM. The specific mathematical expressions are as follows:
(11)
(12)
(13)
where represents the time series input to CNN, represents the features after processing by CNN; and represent the forecasted output results of CNN-LSTM and CNN-Bi-LSTM, respectively. The structure diagrams of CNN-LSTM and CNN-Bi-LSTM are shown in Figure 2.
Figure 2

CNN-LSTM structure diagram, CNN-Bi-LSTM structure only needs to replace the LSTM layer part to Bi-LSTM.

Figure 2

CNN-LSTM structure diagram, CNN-Bi-LSTM structure only needs to replace the LSTM layer part to Bi-LSTM.

Close modal

Structure of the designed CEEMDAN mixed DL model

Complete ensemble empirical mode decomposition adaptive noise

This paper uses the decomposition algorithm CEEMDAN, developed by Torres et al., based on the EEMD algorithm (Torres et al. 2011b). The EMD and EEMD algorithms share the common feature of decomposing the original signal into high-frequency signals called intrinsic mode functions (IMFs) after multiple iterations. EMD and EEMD algorithms can alleviate mode mixing issues by adding positive and negative Gaussian noise to the decomposed signals. This characteristic is also present in CEEMDAN. However, both of these algorithms inevitably leave some white noise in the obtained IMFs from the decomposition of the original signal, which can impact the model's forecasting performance. CEEMDAN, on the other hand, differs in that its key advantage lies in its adaptability. It can automatically adjust noise estimation based on the signal's local characteristics, thereby improving signal decomposition's accuracy and stability. The implementation process of CEEMDAN is as follows:

  • (1). Add white noise with a standard deviation of to the original signal :
    (14)
    where k represents a real number.
  • (2). Decompose the signal set using EMD, then average each decomposition's combined parts.
    (15)
  • (3). Calculate the first-order residuals.
    (16)
  • (4). Further, decompose the signal using EMD and calculate the second IMF mode.
    (17)
    where represents the IMF mode component obtained by EMD.
  • (5). In the next stage, the component and residue are calculated, as given by the following equations:
    (18)
    (19)
  • (6). Repeat Equations (18) and (19) until the residual component no longer satisfies the decomposition conditions, and the decomposition stops. Finally, the original signal can be represented as:
    (20)
    where denotes the final residuals.

CEEMDAN is commonly used for data decomposition and denoising, and its reasonable application to various models is crucial. The specific process is illustrated in Figure 3.
Figure 3

Flowchart of the CEEMDAN-LSTM forecasting process. Note that the LSTM part can be replaced with other models for forecasting.

Figure 3

Flowchart of the CEEMDAN-LSTM forecasting process. Note that the LSTM part can be replaced with other models for forecasting.

Close modal

Ensemble empirical mode decomposition and complete ensemble empirical mode decomposition

To further validate the robustness and complexity analysis of CEEMDAN decomposition, this study proposes a comparison with hybrid ML methods based on EEMD and CEEMD. EEMD is an improvement on the original EMD. EEMD introduces multiple iterations of white noise into the original signal, performs EMD on each noise-added signal, and averages the results. This approach effectively mitigates the mode mixing problem, enhancing the robustness and accuracy of the decomposition (Wu & Huang 2009). CEEMD further refines EEMD. While EEMD reduces mode mixing, residual noise can still occur in some cases. CEEMD addresses this issue by introducing symmetrically distributed positive and negative white noise into the signal, completely eliminating residual noise and ensuring the completeness of the decomposition (Torres et al. 2011a). The forecast structures of hybrid ML models for based on EEMD and CEEMD algorithms are consistent with the structure shown in Figure 3. By replacing the CEEMDAN components with EEMD or CEEMD, the corresponding schematic diagram of the hybrid ML models for prediction is obtained.

CEEMDAN-CNN-LSTM and CEEMDAN-CNN-Bi-LSTM

The hybrid models CEEMDAN-CNN-LSTM and CEEMDAN-CNN-Bi-LSTM involve decomposing the original signal using CEEMDAN. The results are then used as input for CNN to extract features further. The final forecasting is obtained using the features processed by CNN as input for LSTM and Bi-LSTM. For CEEMDAN-CNN-LSTM and CEEMDAN-CNN-Bi-LSTM, the process is shown in Figure 4 and further described as follows.
Figure 4

CEEMDAN-CNN-LSTM structure. Note that LSTM can be replaced by Bi-LSTM that is the structure of CEEMDAN-CNN-Bi-LSTM. Note that replacing CEEMDAN with EEMD or CEEMD results in the corresponding EEMD-CNN-LSTM or CEEMD-CNN-LSTM structure.

Figure 4

CEEMDAN-CNN-LSTM structure. Note that LSTM can be replaced by Bi-LSTM that is the structure of CEEMDAN-CNN-Bi-LSTM. Note that replacing CEEMDAN with EEMD or CEEMD results in the corresponding EEMD-CNN-LSTM or CEEMD-CNN-LSTM structure.

Close modal

Forecasting strategy: ET0 periodic decomposition and sliding window

The forecasting structure adopted in this paper involves using the sliding window algorithm to sum and average daily over multiple days, creating different periods of . The sliding window algorithm has been widely applied in time series tasks for processing (Yahmed et al. 2015; Chou & Ngo 2016; Carmona-Cabezas et al. 2019). It involves setting a fixed-size window as a moving step to construct the time series used as input for the model. In this paper, the forecasting of is done by forecasting the following time series using the previous one. Additionally, the sliding window is used to forecast in each sequence, and the results are compared with the actual values. In this study, daily serves as the foundation. The values for for each period of 3, 5, 7, 10, and 15 days are obtained by adding up the daily values and calculating the respective averages. These averages represent the for periods of 3, 5, 7, 10, and 15 days. The specific forecasting structure is shown in Figure 5. We defined a sliding time window with a length of 1, and the model's input time series is constructed using the sliding window algorithm.
Figure 5

periodic forecast structure.

Figure 5

periodic forecast structure.

Close modal

Tuning of model parameters

This study employed the grid search (GS) hyperparameter optimization method. Since the parameter space of ML and DL algorithms may consist of actual values or have infinite space, GS involves setting multiple values for each parameter within defined bounds. This approach is beneficial for parallelization, making it easy to explore the parameter space and find the optimal model parameters (Liashchynskyi & Liashchynskyi 2019). The ML models, XGBoost, SVM, and RF, have relatively few model parameters, so hyperparameter optimization was not performed on them. XGBoost chose the reg:squarederror parameter, commonly used in regression tasks, as the objective function. The RBF kernel function generally has good generalization ability and is increasingly used in nonlinear mapping SVM (Han et al. 2012). RBF was chosen as the kernel function for SVM. The number of decision trees (N_Estimators) for RF and XGBoost are 100 and 42, respectively. RF requires more decision trees to reduce prediction bias, while XGBoost, with its gradient boosting technique, improves each tree based on the previous one. Therefore, XGBoost uses fewer trees (N_Estimators = 42) for training. In the RF model, random_state serves as the random seed to ensure result reproducibility. In this study, setting it to 50 achieved better forecasting results. In the RF model, setting n_jobs to 1 uses only one computing core, saving computational resources and avoiding interference with other tasks. Setting oob_score = True provides an additional performance evaluation metric, which is useful for tuning model parameters. Setting max_features = 1 ensures that each tree uses all features. min_samples_leaf = 10 specifies that each leaf node must contain at least 10 samples to prevent overfitting. DL models have more parameters. The best model parameters for the DL parameters were determined through GS. In CNN, filters and kernel_size refer to the number of convolutional kernels used for feature extraction and the size of the 1 × 1 kernels, respectively. In LSTM and Bi-LSTM models, the unit parameter represents the number of neurons in the LSTM layer. Higher units increase the model's learning capacity and complexity, but they also lead to greater computational demands and training time, as well as a higher risk of overfitting (Yu et al. 2019). Through multiple experiments and grid searches, this study determined that 150 units is the most suitable choice. Dropout = 0.1 means that each neuron has a 10% chance of being dropped during training, reducing the network's dependence on the training data. A lower dropout value does not affect the model's training stability while helping to prevent overfitting and improve generalization. This study used the Adam optimizer for all DL models. Adam has shown superior performance in model training, as validated by Kingma & Ba (2014). In this study, the forecast is a scalar value, resulting in a single prediction output for the task, so Dense is set to 1. The mean squared error (MSE) was employed as the loss function to monitor the error between the actual values continuously and forecasted values in real-time, aiding in the tuning of model parameters (Kingma & Ba 2014). The batch_size represents the sample size used in each training iteration, while epochs denote the number of iterations for model training. The epochs and batch_size values are the optimal model parameters determined by GS. Suitable epochs and batch_size can enhance the model's prediction speed and stability. The activation function for all DL models is set as a rectified linear unit (ReLU), a nonlinear function with a range from 0 to 1. This function aids the model in learning more complex data relationships. Specific parameters are detailed in Table 2.

Table 2

Specific model parameters for forecasting at the Nanchang, Xinjian, and Dongxiang stations

Station nameModelsParameters
Nanchang, Xinjian, Dongxiang XGB 
  • ▪ Objective = reg:squarederror

  • ▪ n_estimators = 42

 
SVM 
  • ▪ kernel = RBF

 
RF 
  • ▪ n_estimators = 100, oob_score = True

  • ▪ n_jobs = 1, random_state = 50

  • ▪ max_features = 1.0, min_samples_leaf = 10

 
CNN 
  •  filters = 32, kernel_size = 1, activation = ReLU

  •  pool_size = 1, Dropout = 0.1

  •  optimizer = Adam

  •  loss = mse, epochs = 40

 
LSTM 
  •  units = 150, activation = ReLU, Dropout = 0.1

  •  Dense = 1, optimizer = Adam, loss = mse

  •  epochs = 40, batch size = 32

 
Bi-LSTM 
  •  units = 150, activation = ReLU, Dropout = 0.1

  •  Dense = 1, optimizer = Adam, loss = mse

  •  epochs = 30, batch_size = 64

 
CNN-LSTM 
  •  filters = 256, kernel_size = 1, activation = ReLU

  •  Dropout = 0.1, Dense = 1

  •  LSTM (units = 150, activation = ReLU)

  •  Dense = 1, optimizer = Adam, loss = mse

  •  epochs = 30, batch_size = 32

 
CNN-Bi-LSTM 
  •  filters = 256, kernel_size = 1, activation = ReLU

  •  Dropout = 0.1, Dense = 1

  •  Bidirectional (LSTM (units = 150, activation = ReLU)

  •  Dense = 1, optimizer = Adam, loss = mse

  •  epochs = 40, batch_size = 64

 
Station nameModelsParameters
Nanchang, Xinjian, Dongxiang XGB 
  • ▪ Objective = reg:squarederror

  • ▪ n_estimators = 42

 
SVM 
  • ▪ kernel = RBF

 
RF 
  • ▪ n_estimators = 100, oob_score = True

  • ▪ n_jobs = 1, random_state = 50

  • ▪ max_features = 1.0, min_samples_leaf = 10

 
CNN 
  •  filters = 32, kernel_size = 1, activation = ReLU

  •  pool_size = 1, Dropout = 0.1

  •  optimizer = Adam

  •  loss = mse, epochs = 40

 
LSTM 
  •  units = 150, activation = ReLU, Dropout = 0.1

  •  Dense = 1, optimizer = Adam, loss = mse

  •  epochs = 40, batch size = 32

 
Bi-LSTM 
  •  units = 150, activation = ReLU, Dropout = 0.1

  •  Dense = 1, optimizer = Adam, loss = mse

  •  epochs = 30, batch_size = 64

 
CNN-LSTM 
  •  filters = 256, kernel_size = 1, activation = ReLU

  •  Dropout = 0.1, Dense = 1

  •  LSTM (units = 150, activation = ReLU)

  •  Dense = 1, optimizer = Adam, loss = mse

  •  epochs = 30, batch_size = 32

 
CNN-Bi-LSTM 
  •  filters = 256, kernel_size = 1, activation = ReLU

  •  Dropout = 0.1, Dense = 1

  •  Bidirectional (LSTM (units = 150, activation = ReLU)

  •  Dense = 1, optimizer = Adam, loss = mse

  •  epochs = 40, batch_size = 64

 

Evaluation of model performance

The accuracy and performance of the research models in forecasting different periods of were determined using statistical metrics. We used four commonly employed evaluation metrics, along with a combined accuracy (CA) metric that better reflects the stability of the model's predictions, namely, , mean absolute error (),, Nash–Sutcliffe efficiency (NSE), and CA (Adnan et al. 2019), were employed for the evaluation and comparison of the results. Additionally, we used predicted absolute error (PAE) to assess the predictive performance of the model. The specific mathematical expressions are as follows:
(21)
(22)
(23)
(24)
(25)
(26)
where , , and represent the observed, forecasted, and mean values of , respectively. The range of is from 0 to 1, with a higher value indicating higher forecasting accuracy. On the contrary, and have a range from 0 to positive infinity. When the forecasted value is close to the actual value, the numerator gradually approaches 0. In contrast, a larger numerator indicates a greater difference between the forecasted and actual values, leading to larger results. Lower and values represent better forecasting accuracy and performance of the model. NSE and share the same formulaic structure, but the range of NSE is . An NSE value of 1 indicates a perfectly accurate model prediction, while lower values signify poorer predictive performance. The CA value ranges from . A smaller CA value indicates greater stability in the model's forecasting performance. Figure 6 shows a flowchart of the model used in this study.
Figure 6

Model flowchart for forecasting . Note that replacing CEEMDAN in the figure with EEMD and CEEMD results in the corresponding forecast structures.

Figure 6

Model flowchart for forecasting . Note that replacing CEEMDAN in the figure with EEMD and CEEMD results in the corresponding forecast structures.

Close modal

Comparing the forecast accuracy of various models at different periods

To validate the importance of EEMD, CEEMD, and CEEMDAN decomposition algorithms for prediction, an initial comparison was conducted to evaluate the predictive accuracy of various models without using the CEEMDAN decomposition algorithm over different time periods. Tables 35 describe the statistical results of forecasting using three conventional ML models and five DL models in Nanchang, Dongxiang, and Xinjian, respectively. These tables present the statistical analysis of forecasting for the XGBoost (XGB), SVM, RF, CNN, LSTM, Bi-LSTM, CNN-LSTM, and CNN-Bi-LSTM models during the testing phase under six different periods: 1, 3, 5, 7, 10, and 15 days.

Table 3

The statistical values of forecasting for different periods by various models in Nanchang station

 
 

Note: In the table, the cells with a yellow background represent the best forecasting results, while the text in bold indicates the best forecasting results among the three conventional ML models.

Table 4

The statistical values of forecasting for different periods by various models in Dongxiang station

 
 

Note: In the table, the cells with a yellow background represent the best forecasting results, while the text in bold indicates the best forecasting results among the three conventional ML models.

Table 5

The statistical values of forecasting for different periods by various models in Xinjian station

 
 

Note: In the table, the cells with a yellow background represent the best forecasting results, while the text in bold indicates the best forecasting results among the three conventional ML models.

As shown in Table 3, the forecast results vary significantly depending on forecasting models and periods. Compared with forecasting for multiple periods, the forecasting for with a period of 1 day achieved the highest (0.9768) and the lowest , NSE, CA, and , which are 0.2016 , 0.2630 , 0.9768, and 0.161, respectively. The forecasting model that achieved the best results is Bi-LSTM. Compared with several conventional ML models, including XGB, SVM, and RF, the DL-based Bi-LSTM demonstrates higher and lower and in forecasting different periods of . It is worth noting that, in the forecasting of with a period of 10 days, Bi-LSTM's is higher than the model with the lowest forecasting accuracy, XGB, by 16.94%. In addition, the , , , and of Bi-LSTM are respectively 25.83, 37.04, 38.51, and 16.94% higher than those of XGB in terms of forecasting performance. Table 3 shows that, when forecasting for periods of 3, 5, and 7 days, the best-performing model is LSTM. Comparing the results of Bi-LSTM and LSTM for forecasting with periods of 3, 5, and 7 days, the difference in forecasting results is at the level of the third decimal place, indicating a difference of 0.001. Due to the varying patterns of at different periods, the optimal forecasting model does not achieve uniformity. However, with the statistical difference at the 0.001 level, it can be considered an approximate forecasting accuracy in both statistical and practical terms. From Table 5, we can see that when forecasting for different periods, the prediction accuracy of DL algorithms, precisely the LSTM and Bi-LSTM models, is higher than that of ML models such as XGB, SVM, and RF. This indicates that DL algorithms outperform ML algorithms in the forecasting of based on time series. In Table 3, the highest forecasting accuracy is achieved by Bi-LSTM for forecasting for 1 day. When grouped by forecasting period and sorted in descending order of forecasting accuracy, the sequence is as follows: Bi-LSTM forecasting with a period of 1 day, Bi-LSTM forecasting with a period of 10 days, Bi-LSTM forecasting with a period of 15 days, LSTM forecasting with a period of 5 days, LSTM forecasting with a period of 7 days, and LSTM forecasting with a period of 3 days. Tables 4 and 5 show similar trends in forecasting accuracy as Table 3. Table 3 shows that Bi-LSTM has an advantage in forecasting short periods (1 day) of and longer periods (10 and 15 days) of compared with other models. On the other hand, LSTM performs better in forecasting medium periods (3, 5, and 7 days) of compared with other models. This pattern is consistent in both Tables 4 and 5. From the experimental results, Bi-LSTM demonstrates a more stable advantage in forecasting short and long periods of , while LSTM exhibits an advantage in forecasting medium periods of .

Tables 35 show that the forecast results follow similar patterns in different regions. For a detailed analysis based on Nanchang, scatter plots of different values for various periods calculated by the FAO-56 PM model and forecasted by three ML and five DL models are shown in Figure 7. The figure shows that the studied models exhibit significant differences in forecasting for different periods. In forecasting for various periods, the values forecasted by the CNN, LSTM, Bi-LSTM, CNN-LSTM, and CNN-Bi-LSTM models are closer to those calculated by the FAO-56 PM equation. This indicates that models based on DL are more accurate in forecasting , showing a strong correlation with the values obtained from the FAO-56 PM equation. The scatter plot in Figure 7 provides a clearer view of the forecasting accuracy of different models for . In the forecasting of for various periods, the dispersion of data points for XGB, SVM, and RF is significantly higher than that of other models in Figure 7. It can be observed in Figure 7 that when forecasting for a 1-day period, the dispersion of data points for each model is lower. This pattern is further evident in Figure 9. Additionally, in Figure 7, it is clear that for the forecasting of over 3-, 5-, and 7-day periods, the forecasting trend of LSTM is closer to y = x, indicating that its forecasted values are closer to the actual values. However, in the forecasting of over a 15-day period, Bi-LSTM is even closer to the y = x line. This explains that the performance of models varies for different periods. It also indicates that LSTM and Bi-LSTM perform well forecasting for different periods.
Figure 7

Scatter plots of forecasting for different periods by different models at Nanchang station.

Figure 7

Scatter plots of forecasting for different periods by different models at Nanchang station.

Close modal
Figure 8

Three-dimensional curve plots illustrate the forecasting of for 3 (a) and 15 (b) days by LSTM and Bi-LSTM models at Nanchang, Dongxiang, and Xinjian stations.

Figure 8

Three-dimensional curve plots illustrate the forecasting of for 3 (a) and 15 (b) days by LSTM and Bi-LSTM models at Nanchang, Dongxiang, and Xinjian stations.

Close modal
Figure 9

Box plots of absolute forecast errors between the base model and the CEEMDAN model at Nanchang station for periods of 1 day (a) and 15 days (b).

Figure 9

Box plots of absolute forecast errors between the base model and the CEEMDAN model at Nanchang station for periods of 1 day (a) and 15 days (b).

Close modal

To further illustrate its periodicity and the forecasting relationship of LSTM and Bi-LSTM models, Figure 8 provides an analysis of the forecasting results (three-dimensional curve graph) of for different periods in Nanchang, Dongxiang, and Xinjian. It is clear from the graph that for different stations, the two models show significant differences in forecasting for the same period. From Figure 8, it is evident that for forecasting a period of 3 days of , when compared with the curve of the actual values, the forecasting curve of LSTM is closer to the actual values than Bi-LSTM for all three different stations. Conversely, from the graph of the forecasting period of 15 days for , it is clear that the forecasting curve of Bi-LSTM is closer to the actual curve of . Figure 8 effectively illustrates that Bi-LSTM and LSTM have different advantages in forecasting for different periods. Although the difference in their forecasting accuracy is small, Figure 8 suggests that Bi-LSTM is more suitable for forecasting long-period . At the same time, LSTM is more suitable for forecasting shorter periods of . In this study, LSTM and Bi-LSTM accurately forecasted over different periods. Short-term forecasts can provide more precise guidance for daily irrigation, while long-term forecasts contribute to long-term water resource planning.

Comparing the forecasting accuracy of CEEMDAN with the base models at different periods

This study aims to investigate whether adding the CEEMDAN algorithm on top of the base models can lead to higher forecasting accuracy and, consequently, to analyze the importance of CEEMDAN in forecasting . Tables 68 provide a clear overview of the forecasting accuracy of for different stations and periods based on the CEEMDAN algorithm. Analysis of Table 6 reveals that CEEMDAN-Bi-LSTM achieved the best forecasting results when forecasting for a period of 15 days. However, there is no consistent optimal forecasting model for different periods. A different pattern from Tables 35 is observed, wherein the CEEMDAN-based algorithms, forecasting for a 1-day period yielded better accuracy with CEEMDAN-XGB compared with the forecasting accuracy of CEEMDAN-CNN-LSTM, a DL-based model. The same pattern is observed when forecasting the Dongxiang and Xinjian stations. Tables 38 clearly show that in forecasting for different periods, all DL-based algorithms outperform ML-based XGB, SVM, and RF. However, forecasting the for a period of 1 day shows different results, which is in stark contrast to Tables 35. Figure 9 shows the distribution of the optimal models for forecasting for different stations and periods between the base models and those based on CEEMDAN. Analysis of Figure 9 reveals that adding the CEEMDAN algorithm to the forecasting for different periods does not follow a consistent pattern, significantly reducing the stability of the forecasting and the universality of the models.

Table 6

Statistical values of the CEEMDAN model for forecasting ET0 over different periods at the Nanchang station

 
 

Note: In the table, the cells with a yellow background represent the best forecasting results, while the text in bold indicates the best forecasting results among the three conventional ML models.

Table 7

Statistical values of the CEEMDAN model for forecasting ET0 over different periods at the Dongxiang station

 
 

Note: In the table, the cells with a yellow background represent the best forecasting results, while the text in bold indicates the best forecasting results among the three conventional ML models.

Table 8

Statistical values of the CEEMDAN model for forecasting ET0 over different periods at the Xinjian station

 
 

Note: In the table, the cells with a yellow background represent the best forecasting results, while the text in bold indicates the best forecasting results among the three conventional ML models.

Through the analysis of Tables 3 and 6, in the forecasting of with a period of 1 day in Nanchang, the best model of the base model, Bi-LSTM ( = 0.9768, = 0.2016 , = 0.2630 , = 0.9768, CA = 0.1610), outperforms the best model based on CEEMDAN, CEEMDAN-CNN-Bi-LSTM ( = 0.8339, = 0.5206 , = 0.7044 , = 0.8339, CA = 0.4591), by 17.13% in , and is lower in terms of MAE, RMSE, NSE, and CA by 158.23, 167.83, 17.13, and 24.74%, respectively. Tables 68 exhibit similar patterns. Based on Tables 3 and 6, a detailed analysis was conducted. Absolute forecast errors for at Nanchang station with periods of 1 and 15 days were calculated. Box plots were used to visually analyze and compare the absolute forecast errors between the base and CEEMDAN models. We can clearly see from Figure 10 that the base model has fewer outliers in the absolute forecast errors, and, at the same time, the overall values of absolute forecast errors are smaller compared with the CEEMDAN models. This indicates that the base model has higher forecasting accuracy than various CEEMDAN models in forecasting for different periods. To further analyze the forecasting stability between the base model and CEEMDAN, a statistical analysis of the distribution of the best forecast models for all stations and periods of was conducted. From the histogram in Figure 9, it is clear that, compared with the base model, CEEMDAN models exhibit a highly uneven distribution of the best models in forecasting. This indicates that the CEEMDAN models for forecasting are unstable, and the addition of CEEMDAN has yet to improve the generalization of the models. Instead, it has increased the inconsistency of the forecasting models, leading to a decrease in forecast stability.
Figure 10

Distribution of the best models for the forecasting of for different cycles by the base model (a) and the CEEMDAN model (b) at Nanchang, Dongxiang, and Xinjian stations. Note: C-L: CNN-LSTM; C-BL: CNN-Bi-LSTM; CE-XGB: CEEMDAN-XGB; CE-L: CEEMDAN-LSTM; CE-C-L: CEEMDAN-CNN-LSTM.

Figure 10

Distribution of the best models for the forecasting of for different cycles by the base model (a) and the CEEMDAN model (b) at Nanchang, Dongxiang, and Xinjian stations. Note: C-L: CNN-LSTM; C-BL: CNN-Bi-LSTM; CE-XGB: CEEMDAN-XGB; CE-L: CEEMDAN-LSTM; CE-C-L: CEEMDAN-CNN-LSTM.

Close modal

Comparison of the forecasting performance of different decomposition models: EEMD, CEEMD, and CEEMDAN

To further validate the impact and significance of the CEEMDAN decomposition model on multi-period forecasting in humid regions, this study introduced EEMD and CEEMD methods combined with the same ML models as CEEMDAN. The forecasting performance of EEMD and CEEMD-based hybrid ML models was analyzed for multi-period predictions at the Nanchang, Dongxiang, and Xinjian stations. Experimental analysis revealed that EEMD and CEEMD demonstrated similar forecasting patterns and comparable performance across all three stations for multi-period predictions. Therefore, detailed analyses of EEMD and CEEMD forecasting results were conducted based on the results from the Nanchang station. Table 9 presents the model evaluation metrics for the EEMD-based hybrid ML models at Nanchang station over six different periods. Except for predictions at 1- and 3-day periods, EMMD-LSTM and EEMD-Bi-LSTM were identified as the best forecasting models for all other periods, consistent with the results observed for standalone models and CEEMDAN hybrid models. Moreover, Table 9 shows that the EEMD hybrid DL models outperformed the EEMD hybrid traditional ML models in forecasting accuracy.

Table 9

Statistical values of the EEMD model for forecasting over different periods at the Nanchang station

 
 

Note: In the table, the cells with a yellow background represent the best forecasting results, while the text in bold indicates the best forecasting results among the three conventional ML models.

Notably, for predictions at 10- and 15-day forecasting periods, the EEMD-LSTM and EEMD-Bi-LSTM models outperformed EEMD-SVM in terms of R² by approximately 20 and 15.49%, respectively, and in terms of the comprehensive CA metric by approximately 39.47 and 34.37%, respectively. It is worth highlighting that EEMD demonstrated forecasting performance comparable to CEEMDAN across different forecasting periods for . Moreover, in certain forecasting periods, the EEMD method showed superior performance compared with CEEMDAN. However, significant differences remained when compared with the baseline model. For the 10-day forecasting period, Bi-LSTM emerged as the best-performing model. Compared with EEMD-Bi-LSTM, Bi-LSTM achieved superior performance in the comprehensive CA metric, with its forecasting accuracy exceeding that of EEMD-Bi-LSTM by approximately 53.87% under the same forecasting conditions.

In addition to EEMD, this study employed CEEMD combined with the same models as CEEMDAN to analyze the effects of different decomposition methods on predictions across various time periods. Table 10 presents the detailed evaluation metrics for the predictive performance of the CEEMD method. As shown in Table 10, CEEMD-LSTM emerged as the best-performing model for the 5-, 7-, and 10-day forecasting periods, accounting for three out of the six best models across all periods. This result aligns with the predictive patterns observed for the baseline model, the CEEMDAN-based hybrid models, and the EEMD-based hybrid models. LSTM and Bi-LSTM were consistently identified as the most effective models for multi-period prediction in humid regions. Furthermore, hybrid DL models outperformed hybrid traditional ML models, a trend consistent with the results from CEEMDAN and EEMD models. One distinguishing feature of CEEMD compared with EEMD and CEEMDAN was its superior accuracy in predicting for the 15-day forecasting period, where CEEMD-CNN exhibited higher predictive precision. Regarding the comprehensive CA metric, CEEMD-CNN's performance surpassed that of CEEMDAN-CNN and EEMD-CNN by approximately 16.09 and 17.71%, respectively. Nonetheless, the overall predictive performance of CEEMD remained comparable to that of CEEMDAN and EEMD. Compared with the baseline model, CEEMD still demonstrated a noticeable gap in predictive performance for multi-period prediction in humid regions.

Table 10

Statistical values of the CEEMD model for forecasting over different periods at the Nanchang station

 
 

Statistical values of the CEEMDAN model for forecasting ET0 over different periods at the Xinjian station In the table, the cells with a yellow background represent the best forecasting results, while the text in bold indicates the best forecasting results among the three conventional ML models.

This study conducted a detailed boxplot visualization analysis (Figure 11) of the prediction error evaluation metric PAE for forecasts at Nanchang Station over 1- and 15-day forecasting periods. The analysis compared the baseline model (single model), the CEEMDAN-based hybrid model, the EEMD-based hybrid model, and the CEEMD-based hybrid model. From Figure 11(a), it is evident that the prediction errors of the CEEMDAN, EEMD, and CEEMD methods for the 1-day forecast exceed those of the baseline model. Additionally, the three decomposition methods demonstrate similar predictive performance without significant differences. Figure 11(a) also highlights the superior performance of DL methods in multi-period forecasting in humid regions. Similarly, for the 15-day forecast, as shown in Figure 11(b), the predictive performance of the CEEMDAN, EEMD, and CEEMD-based hybrid models is lower than that of the baseline model. Compared with the 1-day forecast, the combination of decomposition methods with DL models (LSTM and Bi-LSTM) shows a more pronounced advantage over traditional ML models (XGB, SVM, and RF) in terms of prediction error. This underscores the superior predictive performance of DL methods.
Figure 11

Box plots of absolute forecast errors between the base model and the CEEMDAN, EEMD, and CEEMD models at Nanchang station for periods of 1 day (a) and 15 days (b).

Figure 11

Box plots of absolute forecast errors between the base model and the CEEMDAN, EEMD, and CEEMD models at Nanchang station for periods of 1 day (a) and 15 days (b).

Close modal

The error analysis presented in Figure 11 clearly illustrates that the CEEMDAN, EEMD, and CEEMD methods exhibit significant performance gaps compared with the baseline model in multi-period forecasting in humid regions. This finding further highlights the influence and limitations of decomposition methods in predicting over multiple periods in humid environments.

Comparison of the computational costs of EEMD, CEEMD, CEEMDAN, and baseline models

Figure 12 illustrates the computational time cost for CEEMDAN models compared with the base models across all meteorological stations. The specific computational cost unit is based on the sum of the time spent on forecasting for all periods at each station. It is evident from Figure 12 that the base CEEMDAN models incur significantly higher computational costs compared with the base models. The figure provides a detailed analysis of the time cost multiples for various models and the runtime of the superior models. The traditional ML models, XGB, SVM, and RF, exhibit extremely fast runtime at each station. This is attributed to the inherent speed of traditional ML, which lacks the complex neural network structures found in DL models. Taking station 58606 (Nanchang station) as a reference, the computational cost of CEEMDAN-based models, LSTM and Bi-LSTM, is approximately 8.9 and 8.5 times higher than that of the base LSTM and Bi-LSTM models, respectively. Similar results are observed at the other two stations. LSTM and Bi-LSTM, compared with CEEMDAN-LSTM and CEEMDAN-Bi-LSTM algorithms, have more superficial structures and lower computational costs while achieving higher forecasting accuracy. Additionally, in terms of the runtime of the EEMD model, the EEMD method required more computation time compared with CEEMDAN. Specifically, during the operation of the EEMD-Bi-LSTM model, the runtime cost was approximately 2.63 times higher than that of CEEMDAN-Bi-LSTM, with a runtime of 11,288 s. Due to the high computational cost, and to minimize the impact on forecast accuracy, this study selected the same model parameters for EEMD-CNN-LSTM and EEMD-CNN-Bi-LSTM as those used in the CEEMDAN-CNN-LSTM and CEEMDAN-CNN-Bi-LSTM models, which reduced the runtime by approximately five times. As shown in Figure 12, CEEMD still incurs significantly higher runtime costs compared with CEEMDAN. Furthermore, the EEMD and CEEMD methods with traditional ML models incur lower computational costs than CEEMDAN. In comparison with the EEMD, CEEMD, and CEEMDAN-LSTM and EEMD, CEEMD, and CEEMDAN-Bi-LSTM algorithms, both LSTM and Bi-LSTM models exhibit a shallower structure and lower computational cost while achieving higher prediction accuracy. As illustrated in Figure 12, the CNN-Bi-LSTM model based on CEEMDAN demonstrates significantly higher runtime costs, highlighting the superior efficiency of LSTM and Bi-LSTM algorithms and the disadvantages of using the EEMD, CEEMD, and CEEMDAN methods.
Figure 12

Three-dimensional histograms of the forecasting time cost for different periods by the base models and EEMD, CEEMD, and CEEMDAN models at Nanchang, Dongxiang, and Xinjian stations. Note that in the figure, CE represents CEEMDA, EM represents EEMD, CEM represents CEEMD, C represents CNN, L represents LSTM, and BL represents Bi-LSTM.

Figure 12

Three-dimensional histograms of the forecasting time cost for different periods by the base models and EEMD, CEEMD, and CEEMDAN models at Nanchang, Dongxiang, and Xinjian stations. Note that in the figure, CE represents CEEMDA, EM represents EEMD, CEM represents CEEMD, C represents CNN, L represents LSTM, and BL represents Bi-LSTM.

Close modal

Strategies for forecasting ET0 in different periods

Accurate forecasting of at different periods is crucial for addressing practical issues, and the model's accuracy varies significantly for different periods. The results of this study indicate that both LSTM and Bi-LSTM models achieve satisfactory results in forecasting at different periods. However, there are some inconsistencies, and as the forecasting period lengthens, the forecasting accuracy tends to decrease. Lu et al. (2023) used VMD-BPNN to forecast at different periods and found that the forecasting accuracy decreases as the forecasting period lengthens. This result is similar to the patterns observed in this study's forecasting of CNN and XGB models. However, in contrast, the forecasting accuracy of LSTM and Bi-LSTM in this study is less affected by the length of the forecasting period and consistently achieves good results across all periods. Multi-step forward forecasting and nonperiodic forecasting of have different meanings. Multi-step forward forecasting allows for advanced knowledge and understanding of the specific data of . Malik et al. (2022) used the MVMD-RR technique to optimize the KELM model for accurate forecasting of for the next 3 and 7 days. However, there is still a phenomenon of decreasing forecasting accuracy as the forecasting length increases. Compared with this experiment, forecasting for different periods allows for a more precise understanding of the specific trends in each period. The results indicate that LSTM and Bi-LSTM do not exhibit a linear decline in forecasting accuracy with changes in the forecasting period length.

Comparison of forecasting accuracy between EEMD, CEEMD, CEEMDAN, and the baseline model CEEMDAN

The basic DL models, LSTM and Bi-LSTM, exhibited similar performance in forecasting for different periods across all stations, outperforming other base models and CEEMDAN models. Mandal & Chanda (2023) found that LSTM had higher forecasting accuracy than SVR, RF, MLP, and MARS in real-time and 28-day ahead forecasting using DL. Ayaz et al. (2021) estimated in two different climatic regions using ML. The evaluation of results indicated that LSTM models outperformed GBR, SVR, and RF models. Long et al. (2022) decomposed the original data through a DWT decomposition model and used LSTM, ANN, and ELM models as base models to estimate in Beijing and Baoding in the northern part of the North China Plain. The results indicated that single ML models (LSTM, ANN, and ELM) were more stable than models based on DWT (wavelet-coupled) decomposition, and they generally exhibited better overall performance. It is worth noting that in a considerable amount of previous research on estimating and forecasting , good results have been obtained using non-CEEMDAN decomposition models. For example, Kang et al. (2022) achieved satisfactory results in forecasting at 10 stations in the Weihe River Basin in China by combining VMD and Box–Cox transformation (BC) with SVM. Jayasinghe et al. (2021) demonstrated the utility of combining MEMD and Boruta with LSTM in forecasting daily datasets, showing its usefulness compared with independent LSTM models. Yan et al. (2023) achieved excellent water-level forecasting results by combining the CEEMDAN decomposition model with LSTM and Bi-LSTM models. The specific analysis of the experimental results above is shown in Table 11. From Table 11, in previous studies, the combination of decomposition algorithms and ML models has yielded satisfactory results for various forecasting tasks. However, in this experiment, the opposite forecasting effect was obtained. In this study, the addition of EEMD, CEEMD, and CEEMDAN led to a decrease in prediction accuracy compared with the base model. The reason for this phenomenon is that the incorporation of EEMD, CEEMD, and CEEMDAN generates components for each feature in the original data. LSTM accepts different component features within a fixed parameter range, greatly increasing the difficulty of generalizing the LSTM model. This significantly increases the dependence on parameters of LSTM and other models, resulting in a significant decrease in forecasting accuracy with parameters that are not suitable for component features, ultimately leading to lower forecasting results than the base model. Additionally, due to the high quality of the data in this experiment, even without using a decomposition algorithm to denoise the raw data, decompose the original signal with higher nonlinearity, and then sequentially forecast each IMF mode, the base model still maintains a high level of forecasting accuracy. The basic LSTM and Bi-LSTM models achieved the best results in forecasting for different periods in this experiment. However, models based on EEMD, CEEMD, and CEEMDAN have lower prediction accuracy than the base models. This is closely related to the data decomposition by EEMD, CEEMD, and CEEMDAN, which introduce changes in input features, requiring the derivation of optimal model parameters corresponding to the altered data. This complexity suggests that EEMD, CEEMD, and CEEMDAN decomposition methods may not be suitable for predicting multi-period forecasts in humid regions .

Table 11

Literature on the combination of decomposition algorithms and ML

Author/ReferencesResearch objectiveBest model usedPerformance of models
Long et al. (2022)  Reference evapotranspiration estimation using long short-term memory network and wavelet-coupled long short-term memory network Wavelet-coupled-LSTM  = 0.998, RMSE = 0.08 , MAE = 0.059 , NSE = 0.997, when Rn, T, RH, U is input. 
Kang et al. (2022)  Machine learning framework with decomposition–transformation and identification of key modes for estimating reference evapotranspiration VMD-Box–Cox (BC)-SVM The correlation coefficient (R) is consistently greater than 0.96; both MAPE (mean absolute percentage error) and RMSE are less than 8.41% and 0.38 mm day−1, respectively. 
Jayasinghe et al. (2021)  Deep multi-stage reference evapotranspiration forecasting model MEMD-Boruta-LSTM R = 0.9668, RMSE = 0.5307 , MAE = 0.4204 , NSE = 0.8960. 
Yan et al. (2023)  Forecasting water levels using CEEMDAN decomposition algorithm with machine learning and deep learning ensemble CEEMDAN-Bi-LSTM  = 0.9846, MAE = 0.501 , RMSE = 0.0641 , nRMSE = 0.0235. 
Author/ReferencesResearch objectiveBest model usedPerformance of models
Long et al. (2022)  Reference evapotranspiration estimation using long short-term memory network and wavelet-coupled long short-term memory network Wavelet-coupled-LSTM  = 0.998, RMSE = 0.08 , MAE = 0.059 , NSE = 0.997, when Rn, T, RH, U is input. 
Kang et al. (2022)  Machine learning framework with decomposition–transformation and identification of key modes for estimating reference evapotranspiration VMD-Box–Cox (BC)-SVM The correlation coefficient (R) is consistently greater than 0.96; both MAPE (mean absolute percentage error) and RMSE are less than 8.41% and 0.38 mm day−1, respectively. 
Jayasinghe et al. (2021)  Deep multi-stage reference evapotranspiration forecasting model MEMD-Boruta-LSTM R = 0.9668, RMSE = 0.5307 , MAE = 0.4204 , NSE = 0.8960. 
Yan et al. (2023)  Forecasting water levels using CEEMDAN decomposition algorithm with machine learning and deep learning ensemble CEEMDAN-Bi-LSTM  = 0.9846, MAE = 0.501 , RMSE = 0.0641 , nRMSE = 0.0235. 

Stability of the EEMD, CEEMD, CEEMDAN, and base models

The stability of the model is an important factor to consider when forecasting reliable . In this study, the three conventional ML models, XGB, SVM, and RF, exhibit the highest percentage variations in , , , , and in forecasting for different periods across all stations. Among them, XGB shows the most noticeable variation in RMSE in forecasting for different periods. This phenomenon also reveals that ML-based XGB, SVM, and RF models, when forecasting different periods of , exhibit a decrease in forecasting performance as the period changes, leading to a reduction in stability. However, for models based on CEEMDAN, judging stability based on the increase in RMSE as a criterion for stability in forecasting different periods of is more stable than the base model. However, for models based on EEMD, CEEMD, and CEEMDAN, the increase in and metrics serves as a criterion for evaluating the stability of predictions across different time periods, revealing that their stability is lower than that of the baseline model. Nevertheless, models based on EEMD, CEEMD, and CEEMDAN should be considered as a cohesive system rather than being evaluated against the stability criteria of the baseline model. These models fail to achieve optimal solutions for different stations and time periods. Figure 10 clearly demonstrates that models based on EEMD, CEEMD, and CEEMDAN demonstrate weaker generalization capabilities, making them incapable of adapting to changes in input features. Consequently, these models produce inconsistent predictions and fail to reliably forecast variations. In contrast, the base models, LSTM and Bi-LSTM, obtained similar forecasting results for at different stations and periods. The percentage increase in RMSE for these models is smaller than for all models in this experiment. Moreover, Figure 12 also shows that the baseline model has the best overall performance compared with the EEMD, CEEMD, and CEEMDAN decomposition models. It also responds more steadily to changes in features, making predictions that are consistent and accurate. Additionally, LSTM and Bi-LSTM based on DL have a large number of parameters that need to be adjusted in real-time based on changes in input features. Therefore, further parameter optimization using GS methods is needed to select the best model parameters to ensure the excellent stability of the models.

Computational costs of the EEMD, CEEMD, CEEMDAN, and base models

The results indicate that the runtime of models based on CEEMDAN is significantly higher than the corresponding base models. Specifically, at the Nanchang station, the computational cost of the CEEMDAN-LSTM model is approximately 8.9 times that of the base LSTM model, and the computational cost of the CEEMDAN-Bi-LSTM model is about 8.5 times that of the base Bi-LSTM model. However, the high multiple of runtime costs comes at the expense of an inversely proportional impact on the final forecasting accuracy. Models based on CEEMDAN exhibit very low operational efficiency. Moreover, the runtime costs for models based on CEEMDAN and CNN are even higher for LSTM and Bi-LSTM. The computational costs of the three conventional ML models, XGB, RF, and SVM, are significantly lower than those of various DL models. However, their forecasting accuracy and stability are lower than those of LSTM and Bi-LSTM. The computational costs of ML increase with the values and quantity of the model parameters used (Fan et al. 2018). DL shares similar principles with ML. LSTM and Bi-LSTM, based on DL, have many parameters, leading to higher computational costs than ML. However, obtaining better forecasting results through higher computational time costs is acceptable. In this experiment, the base models were more efficient in computation time than the CEEMDAN algorithm. Combining high-precision forecasting by LSTM and Bi-LSTM with lower computational time costs than CEEMDAN makes them better models for forecasting . CEEMDAN, which is effective for complex nonlinear problems, decomposes each variable into complex-to-simple signals through multiple iterations and aggregates them to produce a forecast. However, in this study, using CEEMDAN did not improve forecasting performance. This is likely due to the strong correlation within the data itself. For instance, the regional limitations of the forecast area cause significant periodic variations in . This might lead to substantial learning errors when CEEMDAN decomposes signals into simpler components, resulting in inaccurate forecasts. Additionally, decomposing multiple variables with CEEMDAN and reusing them as model inputs results in forecasts equivalent to those based on different datasets, causing a loss of accuracy due to the inconsistency of inputs.

Additionally, we conducted a detailed analysis of the computational time costs for the EEMD and CEEMD models. The results indicate that the EEMD model, when combined with traditional ML methods, demonstrates lower computational time costs compared with the CEEMDAN model. However, this difference is negligible when comparing the computational costs of EEMD and CEEMD combined with DL models. For instance, in the case of EEMD combined with the Bi-LSTM model, the computational cost exceeds 10,000 s, which is 2.6 times higher than that of CEEMDAN-Bi-LSTM. Furthermore, the computational cost of the decomposition methods shows a substantial increase compared with the baseline model. To address this, in this study, EEMD and CEEMD combined with CNN-LSTM and CNN-Bi-LSTM underwent model parameter adjustments. To minimize the impact on forecast accuracy, the model parameters were set to be identical to those of CEEMDAN-CNN-LSTM and CEEMDAN-CNN-Bi-LSTM. Experimental results show that under these parameter settings, computational time was reduced by a factor of five.

However, even without the need for hyperparameter GS algorithms, the computational costs of EEMD and CEEMD models combined with CNN-LSTM and CNN-Bi-LSTM under a single parameter setting still approached half an hour. When using the same GS algorithms as the CEEMDAN and baseline models, the computational time costs for EEMD and CEEMD combined with CNN-LSTM and CNN-Bi-LSTM models remained significantly higher than those of the baseline and CEEMDAN models. This study recommends using LSTM and Bi-LSTM as forecasting models for in future studies of humid regions. Additionally, it provides valuable insights into the use of decomposition models combined with ML for forecasting.

In this study, we compared the performance of LSTM and Bi-LSTM models with other models in forecasting across different periods. The results indicate that the LSTM and Bi-LSTM models demonstrate high accuracy and stability in predicting , consistent with the findings of Mandal & Chanda (2023), who observed that LSTM showed greater forecasting accuracy than SVR, RF, MLP, and MARS in prediction. Additionally, the research by Ayaz et al. (2021) also indicated that the LSTM model outperformed the GBR, SVR, and RF models in estimating in two distinct climate regions. Moreover, a key finding of this study is that EEMD, CEEMD, and CEEMDAN did not outperform the basic LSTM and Bi-LSTM models in multi-period prediction for humid regions of China. This could be due to changes in the correlation between -decomposed IMFs and corresponding IMFs of other meteorological variables after decomposition, potentially weakening these correlations. Furthermore, the temporal dependencies in time-series data may be affected, limiting the model's ability to capture the time-dependent nature of the data accurately, thus leading to a decline in prediction accuracy. It was also observed that these decomposition methods generate multiple IMFs, requiring repetitive predictions for each IMF followed by an aggregation process, which significantly increases the computational cost of the EEMD, CEEMD, and CEEMDAN models. In conclusion, this study highlights the potential of EEMD, CEEMD, and CEEMDAN decomposition methods for multi-period prediction in humid regions of China while emphasizing that prediction should consider not only model accuracy but also computational cost and stability. Due to their superior performance in these aspects, the LSTM and Bi-LSTM models are recommended for multi-period forecasting research in humid regions of China. Our research also provides valuable insights and recommendations for the application of ML combined with decomposition models in forecasting.

In this study, three conventional ML models, XGB, RF, SVM, and DL models, including CNN, LSTM, Bi-LSTM, CNN-LSTM, and CNN-Bi-LSTM, were utilized. In addition, a hybrid model that combines EEMD, CEEMD, and CEEMDAN with the aforementioned models, respectively, was introduced. The objective was to evaluate the accuracy and stability of forecasting different periods of . In the forecasting of for the Nanchang station, the base models, LSTM ( = 0.968, MAE = 0.2245 , RMSE = 0.2686, NSE = 0.968, CA = 0.1727) and Bi-LSTM ( = 0.9768, MAE = 0.2016, RMSE = 0.2630, NSE = 0.9768, CA = 0.1610), achieved satisfactory results. For this station, the best-performing model within the CEEMDAN-based hybrid framework is CEEMDAN-Bi-LSTM (R² = 0.8872, MAE = 0.4853, RMSE = 0.3863 , NSE = 0.8872, CA = 0.3249). Within the EEMD-based hybrid framework, the best-performing model is CEEMD-Bi-LSTM ( = 0.8905, MAE = 0.3606, RMSE = 0.4759 , NSE = 0.8905, CA = 0.3122). Meanwhile, the best-performing model within the CEEMD-based hybrid framework is CEEMD-CNN ( = 0.9038, MAE = 0.3477 , RMSE = 0.4461 , NSE = 0.9038, CA = 0.2937). However, compared with the best-performing model (Bi-LSTM) for the 1-day forecast at the Nanchang station, the EEMD-, CEEMD-, and CEEMDAN-based hybrid Bi-LSTM models exhibited significantly lower forecasting performance in terms of the comprehensive metric CA for the 1-day forecast. Specifically, the prediction performances of the three decomposition models were approximately 193.30, 187.02, and 191.91% lower than that of the baseline Bi-LSTM model, respectively. Furthermore, a statistical analysis of the best models for different stations reveals that the base models consistently obtained a more unified best-performing model for forecasting across various stations and periods. However, the CEEMDAN-based hybrid models exhibited inconsistent best models for forecasting across different stations and periods. This indicates that introducing the CEEMDAN model resulted in poorer forecasting stability. Surprisingly, the analysis of model computational costs revealed an unacceptable increase in the running costs for each CEEMDAN-based hybrid model. The base model Bi-LSTM achieved satisfactory results in forecasting different stations, with an average runtime cost of around 8 min across the three stations. However, after incorporating CEEMDAN, the runtime cost for CEEMDAN-Bi-LSTM increased to approximately 1 h and 10 min. This represents an increase of about 8–9 times in runtime costs after adding CEEMDAN, which is unacceptable. Furthermore, both EEMD and CEEMD demonstrated even higher computational costs compared with CEEMDAN while still showing lower stability than the baseline model, with the computational cost of EEMD-Bi-LSTM being 2.6 times that of CEEMDAN-Bi-LSTM. In conclusion, based on the experimental results, the EEMD, CEEMD, and CEEMDAN models are not suitable for forecasting over different time periods. It is worth noting that incorporating LSTM and Bi-LSTM with CNN is also a decision that requires attention. In this experiment, adding CNN to LSTM and Bi-LSTM did not yield good forecasting results and increased runtime costs. The probable reason for this phenomenon is that the research data in this experiment has high quality, and excessive preprocessing, further feature extraction, and noise reduction methods may have had a counterproductive effect. Similar to enhancing positive relationships but diminishing negative relationships, this could have led to poorer forecasting outcomes. This study investigated the use of EEMD, CEEMD, and CEEMDAN decomposition methods combined with ML to predict in humid regions. The results indicated that the EEMD, CEEMD, and CEEMDAN hybrid models did not outperform the baseline model. It is recommended to use LSTM and Bi-LSTM models for forecasting at different periods in humid regions. Based on the findings, caution should be exercised when using EEMD, CEEMD, or CEEMDAN combined with ML for prediction. Given their accuracy and computational costs, EEMD, CEEMD, or CEEMDAN combined with ML are not optimal choices for forecasting in humid regions and are therefore not recommended for this purpose. Conversely, single models such as XGB, RF, and SVM can be considered for forecasting in humid regions due to their accuracy and time efficiency. It is important to note that this study has some limitations. The integration of EEMD, CEEMD, or CEEMDAN with DL did not show significant improvement in prediction and demonstrated unstable performance across multiple stations and time periods. Additionally, the introduction of EEMD, CEEMD, and CEEMDAN significantly increased the computational cost of the models. In this research, LSTM and Bi-LSTM achieved precise multi-period forecasts for humid regions in China, but they were not validated in other regions, so it cannot be ensured that similar accuracy would be achieved in those areas. In future studies, forecasting for more regions will be conducted to verify the accuracy and stability of LSTM and Bi-LSTM models for multi-period forecasting across different regions.

We would like to express our special thanks to the National Meteorological Information Center of China Meteorological Administration for providing the foundational data for this research.

Conceptualization, Z.Y. and L.W.; methodology, Z.Y. and L.W.; software, Z.Y.; validation, Z.Y., L.W., and X.L.; formal analysis, Z.Y. and L.W.; investigation, Z.Y., L.W., and X.L.; resources, Z.Y. and L.W.; data curation, Z.Y.; writing – original draft preparation, Z.Y.; writing – review and editing, Z.Y. and L.W.; visualization, Z.Y. and L.W.; supervision, Z.Y., L.W., and X.L.; project administration, X.L.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

This research was funded by the General Project of Natural Science Foundation of Jiangxi Province (Grant No. 20232BAB205031), the National Natural Science Foundation of China (Grant No. 52269013), and the Key Project of Natural Science Foundation of Jiangxi Province (Grant No. 20242BAB26081).

Data cannot be made publicly available; readers should contact the corresponding author for details.

The authors declare there is no conflict.

Abdullah
S. S.
,
Malek
M. A.
,
Abdullah
N. S.
,
Kisi
O.
&
Yap
K. S
. (
2015
)
Extreme Learning Machines: A new approach for prediction of reference evapotranspiration
.
Journal of Hydrology
,
527
,
184
195
.
Adnan
R. M.
,
Liang
Z.
,
Trajkovic
S.
,
Zounemat-Kermani
M.
,
Li
B.
&
Kisi
O
. (
2019
)
Daily streamflow prediction using optimally pruned extreme learning machine
.
Journal of Hydrology
,
577
,
123981
.
Ali
M.
,
Jamei
M.
,
Prasad
R.
,
Karbasi
M.
,
Xiang
Y.
,
Cai
B.
,
Abdulla
S.
,
Ahsan Farooque
A.
&
Labban
A. H
. (
2023
)
New achievements on daily reference evapotranspiration forecasting: Potential assessment of multivariate signal decomposition schemes
.
Ecological Indicators
,
155
,
111030
.
Allen
R. G.
,
Pereira
L. S.
,
Raes
D.
&
Smith
M
. (
1998
)
Crop evapotranspiration-Guidelines for computing crop water requirements-FAO Irrigation and drainage paper 56. Fao
,
Rome
,
300
(
9
),
D05109
.
Ayaz
A.
,
Rajesh
M.
,
Singh
S. K.
&
Rehana
S
. (
2021
)
Estimation of reference evapotranspiration using machine learning models with limited data
.
AIMS Geosci
,
7
(
3
),
268
290
.
Bickici Arikan
B.
,
Jiechen
L.
,
I D Sabbah
I.
,
Ewees
A.
,
Homsi
R.
&
Sulaiman
S. O.
(
2021
)
Dew Point Time Series Forecasting at the North Dakota
.
Knowledge-Based Engineering and Sciences
,
2
(
2
),
24
34
.
Breiman
L.
(
2001
)
Random forests
,
Machine Learning
,
45
,
5
32
.
Cao
K.
,
Kim
H.
,
Hwang
C.
&
Jung
H
. (
2018
)
CNN-LSTM coupled model for prediction of waterworks operation data
.
Journal of information processing systems
,
14
(
6
),
1508
1520
.
Carmona-Cabezas
R.
,
Gómez-Gómez
J.
,
Gutiérrez de Ravé
E.
&
& Jiménez-Hornero
F. J.
(
2019
)
A sliding window-based algorithm for faster transformation of time series into complex networks
.
Chaos: An Interdisciplinary Journal of Nonlinear Science
,
29
(
10
),
103121
.
Cetin
M.
,
Alsenjar
O.
,
Aksu
H.
,
Golpinar
M. S.
&
Akgul
M. A
. (
2023
)
Comparing actual evapotranspiration estimations by METRIC to in-situ water balance measurements over an irrigated field in Turkey
.
Hydrological Sciences Journal
,
68
(
8
),
1162
1183
.
Chen
T.
&
Guestrin
C.
(
2016
) '
XGBoost: a scalable tree boosting system
',
Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
, pp.
785
794
.
Chen
H.
,
Huang
J. J.
,
Dash
S. S.
,
McBean
E.
,
Singh
V. P.
,
Li
H.
,
Wei
Y.
,
Zhang
P.
&
Zhou
Z
. (
2023
)
A non-linear theoretical dry/wet boundary-based two-source trapezoid model for estimation of land surface evapotranspiration
.
Hydrological Sciences Journal
,
68
(
11
),
1591
1609
.
Cortes
C.
&
Vapnik
V.
(
1995
)
Support-vector networks
,
Machine Learning
,
20
,
273
297
.
Du
S.
,
Li
T.
,
Yang
Y.
&
Horng
S. J
. (
2021
)
Deep Air Quality Forecasting Using Hybrid Deep Learning Framework
.
IEEE Transactions on Knowledge and Data Engineering
,
33
(
6
),
2412
2424
.
Feng
Y.
,
Cui
N.
,
Gong
D.
,
Zhang
Q.
&
Zhao
L
. (
2017a
)
Evaluation of random forests and generalized regression neural networks for daily reference evapotranspiration modelling
.
Agricultural Water Management
,
193
,
163
173
.
Feng
Y.
,
Peng
Y.
,
Cui
N.
,
Gong
D.
&
Zhang
K
. (
2017b
)
Modeling reference evapotranspiration using extreme learning machine and generalized regression neural network only with temperature data
.
Computers and Electronics in Agriculture
,
136
,
71
78
.
Ferreira
L. B.
&
da Cunha
F. F.
(
2020a
)
Multi-step ahead forecasting of daily reference evapotranspiration using deep learning
.
Computers and Electronics in Agriculture
,
178
,
105728
.
Ferreira
L. B.
,
da Cunha
F. F.
,
de Oliveira
R. A.
&
Fernandes Filho
E. I.
(
2019
)
Estimation of reference evapotranspiration in Brazil with limited meteorological data using ANN and SVM – A new approach
.
Journal of Hydrology
,
572
,
556
570
.
Gao
B.
,
Huang
X.
,
Shi
J.
,
Tai
Y.
&
Zhang
J
. (
2020
)
Hourly forecasting of solar irradiance based on CEEMDAN and multi-strategy CNN-LSTM neural networks
.
Renewable Energy
,
162
,
1665
1683
.
Gocic
M.
,
Petković
D.
,
Shamshirband
S.
&
Kamsin
A
. (
2016
)
Comparative analysis of reference evapotranspiration equations modelling by extreme learning machine
.
Computers and Electronics in Agriculture
,
127
,
56
63
.
Han
S.
,
Cao
Q.
&
Han
M.
(
2012
) '
Parameter selection in SVM with RBF kernel function
',
World Automation Congress 2012
,
24–28 June 2012
, pp.
1
4
.
Hargreaves
G. H.
&
Samani
Z. A.
(
1985
)
Reference crop evapotranspiration from temperature
,
Applied Engineering in Agriculture
,
1
(
2
),
96
99
.
Hochreiter
S.
&
Schmidhuber
J.
(
1997
)
Long short-term memory
,
Neural Computation
,
9
(
8
),
1735
1780
.
Hoseinzade
E.
&
Haratizadeh
S.
(
2019
)
CNNpred: CNN-based stock market prediction using a diverse set of variables
,
Expert Systems with Applications
,
129
,
273
285
.
Ikram
R. M. A.
,
Mostafa
R. R.
,
Chen
Z.
,
Parmar
K. S.
,
Kisi
O.
&
Zounemat-Kermani
M
. (
2023
)
Water Temperature Prediction Using Improved Deep Learning Methods through Reptile Search Algorithm and Weighted Mean of Vectors Optimizer
.
Journal of Marine Science and Engineering
,
11
(
2
),
259
.
Janiesch
C.
,
Zschech
P.
&
Heinrich
K.
(
2021
)
Machine learning and deep learning
,
Electronic Markets
,
31
(
3
),
685
695
.
Jensen
M. C.
(
1968
)
The performance of mutual funds in the period 1945–1964
,
The Journal of Finance
,
23
(
2
),
389
416
.
Jerin
J. N.
,
Islam
H. M. T.
,
Islam
A. R. M. T.
,
Shahid
S.
,
Hu
Z.
,
Badhan
M. A.
,
Chu
R.
&
Elbeltagi
A
. (
2021
)
Spatiotemporal trends in reference evapotranspiration and its driving factors in Bangladesh
.
Theoretical and Applied Climatology
,
144
(
1
),
793
808
.
Karbasi
M.
,
Ali
M.
,
Randhawa
G. S.
,
Jamei
M.
,
Malik
A.
,
Shah
S. H. H.
,
Bos
M.
,
Zaman
Q.
&
Farooque
A. A
. (
2024
)
Innovative multi-temporal evapotranspiration forecasting using empirical fourier decomposition and bidirectional long short-term memory
.
Smart Agricultural Technology
,
9
,
100619
.
Kingma
D. P.
&
Ba
J.
(
2014
)
Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980
.
Kişi
Ö
. (
2006
)
Generalized regression neural networks for evapotranspiration modelling
,
Hydrological Sciences Journal
,
51
(
6
),
1092
1105
.
Liashchynskyi
P.
&
Liashchynskyi
P.
(
2019
)
Grid search, random search, genetic algorithm: a big comparison for NAS. arXiv preprint arXiv:1912.06059
.
Liaw
A.
&
Wiener
M.
(
2002
)
Classification and regression by randomForest
,
R News
,
2
(
3
),
18
22
.
Luo
Y.
,
Chang
X.
,
Peng
S.
,
Khan
S.
,
Wang
W.
,
Zheng
Q.
&
Cai
X
. (
2014
)
Short-term forecasting of daily reference evapotranspiration using the Hargreaves–Samani model and temperature forecasts
.
Agricultural Water Management
,
136
,
42
51
.
Mallat
S. G.
(
1989
)
A theory for multiresolution signal decomposition: the wavelet representation
,
IEEE Transactions on Pattern Analysis and Machine Intelligence
,
11
(
7
),
674
693
.
Mehdizadeh
S.
,
Behmanesh
J.
&
Khalili
K.
(
2017
)
Using MARS, SVM, GEP and empirical equations for estimation of monthly mean reference evapotranspiration
,
Computers and Electronics in Agriculture
,
139
,
103
114
.
Pal
M.
&
Deswal
S.
(
2009
)
M5 model tree based modelling of reference evapotranspiration
,
Hydrological Processes: An International Journal
,
23
(
10
),
1437
1443
.
Pendiuk
J. E.
,
Guarracino
L.
,
Reich
M.
&
Güntner
A
. (
2023
)
Estimating cumulative evapotranspiration using superconducting gravimeter data: a study in Buenos Aires Province, Argentina
.
Hydrological Sciences Journal
,
68
(
15
),
2262
2275
.
Perera
K. C.
,
Western
A. W.
,
Nawarathna
B.
&
George
B
. (
2014
)
Forecasting daily reference evapotranspiration for Australia using numerical weather prediction outputs
.
Agricultural and forest meteorology
,
194
,
50
63
.
Priestley
C. H. B.
&
Taylor
R. J.
(
1972
)
On the assessment of surface heat flux and evaporation using large-scale parameters
,
Monthly Weather Review
,
100
(
2
),
81
92
.
Sabanci
D.
,
Yurekli
K.
,
Comert
M. M.
,
Kilicarslan
S.
&
Erdogan
M
. (
2023
)
Predicting reference evapotranspiration based on hydro-climatic variables: comparison of different machine learning models
.
Hydrological Sciences Journal
,
68
(
7
),
1050
1063
.
Saggi
M. K.
&
Jain
S.
(
2019
)
Reference evapotranspiration estimation and modeling of the Punjab Northern India using deep learning
,
Computers and Electronics in Agriculture
,
156
,
387
398
.
Schuster
M.
&
Paliwal
K. K.
(
1997
)
Bidirectional recurrent neural networks
,
IEEE Transactions on Signal Processing
,
45
(
11
),
2673
2681
.
Shi
H.
,
Xu
M.
&
Li
R.
(
2018
)
Deep learning for household load forecasting – a novel pooling deep RNN
,
IEEE Transactions on Smart Grid
,
9
(
5
),
5271
5280
.
Soman
K.
,
Poornachandran
P.
,
Athira
S.
&
Harikumar
K
. (
2015
)
Recursive variational mode decomposition algorithm for real time power signal decomposition
.
Procedia Technology
,
21
,
540
546
.
Tao
H.
,
Diop
L.
,
Bodian
A.
,
Djaman
K.
,
Ndiaye
P. M.
&
Yaseen
Z. M
. (
2018
)
Reference evapotranspiration prediction using hybridized fuzzy model with firefly algorithm: Regional case study in Burkina Faso
.
Agricultural Water Management
,
208
,
140
151
.
Tikhamarine
Y.
,
Malik
A.
,
Kumar
A.
,
Souag-Gamane
D.
&
Kisi
O
. (
2019
)
Estimation of monthly reference evapotranspiration using novel hybrid machine learning approaches
.
Hydrological sciences journal
,
64
(
15
),
1824
1842
.
Torres
A. F.
,
Walker
W. R.
&
Mckee
M.
(
2011a
)
Forecasting daily potential evapotranspiration using machine learning and limited climatic data
,
Agricultural Water Management
,
98
(
4
),
553
562
.
Torres
M. E.
,
Colominas
M. A.
&
Schlotthauer
G.
& Flandrin,
P
. (
2011b
)
A complete ensemble empirical mode decomposition with adaptive noise, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 22–27 May 2011
.
Trabert
W.
(
1896
)
Neue beobachtungen über verdampfungsgeschwindigkeiten
,
Meteorologische Zeitschrift
,
13
,
261
263
.
Trajkovic
S.
(
2005
)
Temperature-based approaches for estimating reference evapotranspiration
,
Journal of Irrigation and Drainage Engineering
,
131
(
4
),
316
323
.
Viton
F.
,
Elbattah
M.
,
Guérin
J.-L.
&
Dequen
G.
(
2020
)
Heatmaps for visual explainability of CNN-based predictions for multivariate time series with application to healthcare, 2020 IEEE International Conference on Healthcare Informatics (ICHI), pp.
1
8
.
Vosoughifar
H.
,
Khoshkam
H.
,
Bateni
S. M.
,
Jun
C.
,
Xu
T.
,
Band
S. S.
&
Neale
C. M. U
. (
2023
)
Estimation of daily reference evapotranspiration from limited climatic variables in coastal regions
.
Hydrological Sciences Journal
,
68
(
1
),
91
107
.
Wang
L.
,
Kisi
O.
,
Zounemat-Kermani
M.
&
Li
H
. (
2017
)
Pan evaporation modeling using six different heuristic computing methods in different climates of China
.
Journal of Hydrology
,
544
,
407
427
.
Wu
Z.
&
Huang
N. E.
(
2009
)
Ensemble empirical mode decomposition: a noise-assisted data analysis method
,
Advances in Adaptive Data Analysis
,
01
(
01
),
1
41
.
Yahmed
Y. B.
,
Bakar
A. A.
,
RazakHamdan
A.
,
Ahmed
A.
&
Syed Abdullah
S. M
. (
2015
)
Adaptive sliding window algorithm for weather data segmentation
,
Journal of Theoretical & Applied Information Technology
,
80
(
2
),
322
333
.
Yeh
J.-R.
,
Shieh
J.-S.
&
Huang
N. E.
(
2010
)
Complementary ensemble empirical mode decomposition: a novel noise enhanced data analysis method
,
Advances in Adaptive Data Analysis
,
2
(
02
),
135
156
.
Yu
Y.
,
Si
X.
,
Hu
C.
&
Zhang
J
. (
2019
)
A review of recurrent neural networks: LSTM cells and network architectures
.
Neural computation
,
31
(
7
),
1235
1270
.
Yuan
X.
,
Chen
C.
,
Lei
X.
,
Yuan
Y.
&
Muhammad Adnan
R
. (
2018
)
Monthly runoff forecasting based on LSTM–ALO model
.
Stochastic Environmental Research and Risk Assessment
,
32
(
8
),
2199
2212
.
Zhang
W.
,
Qu
Z.
,
Zhang
K.
,
Mao
W.
,
Ma
Y.
&
Fan
X
. (
2017
)
A combined model based on CEEMDAN and modified flower pollination algorithm for wind speed forecasting
.
Energy Conversion and Management
,
136
,
439
451
.
Zhang
Z.
,
Gong
Y.
&
Wang
Z.
(
2018
)
Accessible remote sensing data based reference evapotranspiration estimation modelling
,
Agricultural Water Management
,
210
,
59
69
.
Zhang
L.
,
Zhao
X.
,
Zhu
G.
,
He
J.
,
Chen
J.
,
Chen
Z.
,
Traore
S.
,
Liu
J.
&
Singh
V. P
. (
2023
)
Short-term daily reference evapotranspiration forecasting using temperature-based deep learning models in different climate zones in China
.
Agricultural Water Management
,
289
,
108498
.
Zhao
X.
,
Li
Y.
,
Zhao
Z.
,
Xing
X.
,
Feng
G.
,
Bai
J.
,
Wang
Y.
,
Qiu
Z.
&
Zhang
J
. (
2022
)
Prediction Model for Daily Reference Crop Evapotranspiration Based on Hybrid Algorithm in Semi-Arid Regions of China
.
Atmosphere
,
13
(
6
),
922
.
Zheng
Z.
,
Ali
M.
,
Jamei
M.
,
Xiang
Y.
,
Karbasi
M.
,
Yaseen
Z. M.
&
Farooque
A. A
. (
2023
)
Design data decomposition-based reference evapotranspiration forecasting model: A soft feature filter based deep learning driven approach
.
Engineering Applications of Artificial Intelligence
,
121
,
105984
.
This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (CC BY 4.0), which permits copying, adaptation and redistribution, provided the original work is properly cited (http://creativecommons.org/licenses/by/4.0/).