In drought-prone regions like Australia, accurately assessing evaporation rates is essential for effectively managing and maximising the use of precious water resources and reservoirs. Current estimates show that evaporation reduces Australia's open water lake capacity by about 40% annually. With climate change, this water loss is expected to become an even greater concern. This study investigates a transformer-based neural network (TNN) to estimate monthly evaporation in three Australian locations. The models were trained and tested using monthly weather data spanning from 2009 to 2022. Input parameters were chosen based on Pearson's correlation coefficient values to identify the most impactful combinations. The developed TNN model was compared with two widely used empirical methods, namely Thornthwaite and Stephens and Stewart. The TNN model's impressive accuracy in evaporation prediction, attributed to its unique self-attention mechanism, suggests its promising potential for future use in evaporation forecasting. Additionally, the study revealed an intriguing result: Despite using the same input datasets, the TNN model surpassed traditional methods, achieving an average improvement of 18% in prediction accuracy. The TNN prediction model accurately predicts water loss (average R² = 0.970), supports irrigation management and agricultural planning and offers financial benefits to farming and related industries.

  • A self-attention mechanism based on transformers has been developed to forecast pan evaporation.

  • Utilise meteorological data to anticipate evaporation rates in three distinct Australian regions.

  • Precise forecasts of evaporation can enhance water resource management practices.

  • The transformer model stands out as an effective tool for predicting water loss attributed to evaporation.

AI

artificial intelligence

ANN

artificial neural network

DL

deep learning

Ep

pan evaporation

MAE

mean absolute error

ML

machine learning

NSE

Nash–Sutcliffe efficiency

PCC

Pearson correlation coefficient

R2

coefficient of determination

RH

relative humidity

RMSE

root mean square error

Rs

solar radiation

SS

Stephens and Stewart

WS

wind speed

Ta

mean air temperature

TH

Thornthwaite

Tmax

maximum air temperature

Tmin

minimum air temperature

TNN

transformer neural network

Background

Anticipating evaporation is essential for efficient irrigation and water management, as well as for optimising water usage and agricultural forecasting. This is because evaporation is among the primary drivers of the hydrological cycle. Furthermore, an increased rate of evaporation is a sign of rising temperatures (Chen et al. 2018). Thus, monitoring and managing water resources necessitates vigilant observation of evaporation patterns (Jasmine et al. 2022). The evaporation leads to a significant decrease in levels of water in lakes, reservoirs, and overall, in the budget for water. Thus, evaporation losses must be projected prior to putting water resource policies into practice and planning watering systems. Evaporation rates are influenced by the vapour pressure differentials and heat availability; these decisive elements are influenced by meteorological elements such as humidity, solar radiation, wind speed, air pressure, and air temperature (Fan et al. 2018). Other criteria, including location, kind of climate, seasonal effect, and time of day, are also strongly correlated with such aspects. As a result, evaporation is a complicated phenomenon with highly non-linear properties (Abed et al. 2022a).

Evaporation is projected using direct as well as indirect techniques, such as the evaporation pan, Penman approach, water balance, energy balance, and mass transfer (Wu et al. 2020a). Kisi et al. (2016) showed that the widely favoured approach is the use of the evaporation pan method, chosen for its cost-effectiveness and simplicity. Additionally, this method offers an accurate evaluation of variations in evaporation (Kahler & Brutsaert 2006). The present work aims to project an evaporation pan (Ep) with a precision comparable to real evaporation. On the other hand, approaches based on weather data linked to the energy budget, water budget, and experimental evaporation equation have been employed for Ep estimation (Wang et al. 2016). For instance, in McJannet et al. (2008), a survey of existing literature and methods to gauge evaporation pinpointed ‘combination methods’ as the most suitable for calculating evaporation from open water in an Australian locale using the current data. The chosen combination method involved applying the Penman–Monteith technique, with a modification accounting for variations in heat storage within the water body to adjust the available energy for evaporation. This method has been used and tested in other countries, in Bontempo Scavo et al. (2021), it has been tested in Italy, and further experimental expressions have been tested. However, the linear modelling method does not adequately capture the subtle stochastic aspects of the evaporative process, which could result in very significant inaccuracies (Abed et al. 2021a). Furthermore, due to the inherent variability in empirical models' performance across various circumstances, the calibration of their parameters becomes indispensable to tailoring them to specific zones, thereby introducing an additional layer of complexity to their utilisation. Due to insufficient performance levels, these methods have spurred scientists to explore alternative approaches like artificial intelligence (AI) for accurately assessing evaporation levels (Abed et al. 2022b). In recent decades, AI has demonstrated its effectiveness in addressing numerous complex engineering challenges by leveraging its capacity to perform tasks that traditionally necessitate human intelligence. These tasks include problem-solving, recognising patterns, and making decisions, often at scales and speeds far exceeding human capabilities (Haykin & Lippmann 1994).

Literature review

Various water engineering and environmental concerns have been addressed using AI techniques including artificial neural network (ANN), extreme learning machine (ELM), adaptive neuro-fuzzy inference system (ANFIS), and gene expression programming (GEP) (Khairan et al. 2022). These AI methods are more user-friendly, dependable, and capable of faithfully mimicking complex non-linear processes (Abed et al. 2010). Research into applying AI to calculate various hydrological metrics has been extensively studied (Al-Mukhtar 2021). Scientists have determined that ANNs are more precise in their estimations than those generated using conventional methods (Ditthakit et al. 2022). As a result, several engineering research domains have successfully used AI-based modelling tools.

Through the application of AI techniques and various optimisation methods, numerous studies have explored predicting pan evaporation rates, considering the challenges associated with both practical and theoretical measurement methodologies mentioned earlier (Kumar et al. 2021; Ikram et al. 2022; Kisi et al. 2022; Adnan et al. 2023). In fact, a majority of these works primarily emphasised determining the generalised capabilities of AI frameworks for different weathers since climatic features are stochastic and non-stationary (Abed et al. 2023b). These works provided specific machine learning (ML) models to address the problem based on varying input sets comprising prevalent climatic factors like temperature, wind velocity, humidity, vapour pressure, sunshine, and solar radiation (Abed et al. 2021b). For example, Keskin & Terzi (2006) employed ANN and Penman frameworks and devised a predictive evaporation modelling system. They used numerous meteorological factors as ML inputs. These researchers suggested that ML had a better performance than the Penman method for Ep prediction. Kişi (2013) devised evolutionary neural networks to forecast monthly Ep. The results suggested that the developed approaches offered excellent accuracy compared with empirical methods. Deo et al. (2016) assessed monthly evaporation and associated water loss; they employed three ML approaches: multivariate adaptive regression spline (MARS), ELM, and relevance vector machine (RVM). RVM was found to be the most successful out of the three approaches when using meteorological variables as predictor variables. Falamarzi et al. (2014) assessed the ability of ANN and wavelet-based ANNs to predict evaporation on a daily scale based on temperature and wind speed as input data. The finding suggested that these two approaches predicted evaporation accurately. Reports indicate AI learning techniques' superiority at estimating Ep for different climatic scenarios. Specifically, deep learning (DL) algorithms have been extensively employed across various engineering research fields to create precise and dependable predictions (Arif et al. 2022). DL is a subset of ML using advanced neural networks to analyse and learn from large datasets, enabling machines to recognise patterns, make predictions, and perform complex tasks autonomously.

Considering that the DL techniques that employ enhanced neural networks with multiple layers are suitable for data from time series, they enable advanced approaches to estimate Ep. Their effectiveness can be attributed to their presently rising popularity concerning AI approaches employed for business and scientific domains considering increased accuracy (Hu et al. 2018). Recurrent neural networks (RNNs), a foundational element of DL techniques, are particularly well-suited for predicting and analysing time series data since they can use and store memory from previous network states, a noteworthy capability (Chang et al. 2016). Despite a conventional RNN model's ability to process time series data trends, retaining the variables' extended dependence is a concern, like exploding and vanishing gradients (Bengio et al. 1994). Considering these two critical challenges associated with a typical RNN, network training could produce zero or impractically large network weights. In practical terms, for improved network training, it is critical to retain essential information and omit redundant or unneeded data regarding various network states.

Recently, researchers have effectively employed attention-based approaches for time series estimation. The self-attention (intra-attention) phenomenon is the foundation of the transformer system that has been in widespread use recently (Shao et al. 2019). Transformers were initially employed for machine transliteration applications by Vaswani et al. (2017). Researchers indicated a noteworthy ability to generalise critical tasks like sequence processing and computer vision. A transformer network does not suffer from the vanishing gradient challenge, which is a common issue for recurrent networks. Furthermore, it can utilise all past data points, regardless of the temporal distance between them, unlike recurrent networks. The transformer variant enables parallel processing, allowing for higher speed than recurrent networks. This capability gives the transformer the capacity to identify long-term dependencies without needing to go through sequential processing. Overall, transformer networks do not assess inputs sequentially because the system depends on the self-attention phenomenon that surmounts specific challenges concerning recurrent and convolutional sequence-to-sequence approaches (Wang et al. 2021). Apart from speed and efficiency, the transformer offers the benefit of providing better interpretability compared with other approaches. Contrary to RNNs that might have interpretation difficulties owing to sequential processing, transformer outputs are simpler and transparent. Therefore, it is relatively easy to comprehend model predictions and regulate as required. Researchers have successfully used transformer models for various tasks concerning time series estimation, which is superior to many prediction methods. Further studies have been performed to enhance recurrent DL networks by incorporating self-attention mechanisms. For instance, a deep transformer approach employed to estimate illness resembling influenza was devised by Wu et al. (2020b), and it is superior to Long Short-Term Memory (LSTM) and sequence-to-sequence approaches. The self-attention feature of the transformer model provides better estimation performance compared with the sequence-to-sequence systems using linear attention. Given its strength to handle the intricacies of time series data, the use of transformers in simulations has the potential to be incredibly powerful. Sequence approaches, on the other hand, face significant difficulty in replicating the complex dynamics of such data. The vanishing gradient problem, which prevents RNNs from being able to effectively forecast over longer time frames, can potentially be addressed in a comprehensive manner by using this technique.

The literature validated that employing AI along with appropriate algorithms can precisely simulate evaporation for many regions and offer excellent outcomes compared with comparatively sophisticated conventional techniques (Lu et al. 2018). However, determining and devising a trustworthy, efficient, and generalised prediction approach remains difficult owing to the complex non-linear phenomena with evaporation. Of the numerous ML techniques utilised in recent times, the DL frameworks exhibited wide perspective towards addressing prediction challenges. It is known to provide superior outcomes to relatively sophisticated methods. Specifically, the literature indicates that among several DL models, attention-based models were employed similarly for estimating time series data with a noteworthy ability to address the challenges concerning convolutional sequence-to-sequence and RNN approaches. This study offers a potent approach concerning evaporative losses by employing a transformer model with the self-attention mechanism to estimate Ep in Australia. Successful development would be salutary for managing water resources to allow sustainable farming.

Objectives

The aim of this study is to investigate the use, predictability, and precision of transformer neural network (TNN) approaches for processing climatological data from 2009 to 2022 and estimating monthly Ep magnitudes for three regions in Australia. The predictive precision of the technique is evaluated through several input patterns to obtain maximum precision. The performance of the TNN technique is compared with two common empirical methods, namely Thornthwaite and Stephens and Stewart, based on the same input data. Model efficiency metrics are processed and assessed based on statistical performance indicators deemed appropriate for estimating evaporation rates. In addition, an adequate assessment was performed in this study to establish that TNN modelling is a reliable approach for predicting evaporation, a process that is especially important for agricultural management and water resource management.

Study area

Like most semi-arid regions, Australia depends on stored water reservoirs to meet food production and drinking water needs. Unfortunately, evaporation rates in these countries could be extremely high. Australia loses about 40% of its overall stored water annually owing to high evaporation rates (Craig et al. 2005). Open water evaporation in Australia is a growing concern due to the problems created by the impacts of climate change. Therefore, devising a precise prediction approach to estimate the water deficit is strategic for creating hydrological and water resource planning approaches in this drought-affected country. The present study used data from three areas located across different regions in Australia (Figure 1): Alice Springs (longitude 133°52′50″ E, latitude 23°41′52″ S, elevation 545 m), Brisbane (longitude 153°1′15.85″ E, latitude 27°28′12.45″ S, elevation 4 m), and Adelaide (longitude 138°35′55.07″ E, latitude 34°55′43.18″ S, elevation 6 m). The chosen study areas provide a distinctive opportunity to evaluate AI-driven predictive models due to their unique geographical locations and varied climatic conditions. One notable example is Alice Springs, situated in a desert climate zone characterised by semi-arid conditions featuring hot summers. In contrast, Adelaide, the capital of South Australia, experiences a Mediterranean climate with hot, dry summers and mild, rainy winters. Brisbane, the capital of Queensland, has a subtropical climate marked by hot, humid summers and mild, dry winters, with rainfall evenly distributed throughout the year. January and February typically record the highest rainfall, while July and August are usually the driest months. Additionally, Brisbane encounters occasional thunderstorms, particularly in the summer season. Moreover, the chosen sites are large cities that support large populations. Also, these areas comprise relatively huge land parcels for farm operations, specifically to produce a wide variety of agricultural products. Hence, this work aims to create a DL framework to estimate evaporation. It is critical for managing water resources and practising sustainable agriculture.
Figure 1

Case study locations.

Figure 1

Case study locations.

Close modal

Data description

Monthly data covering the period from 1 January 2009 to 31 October 2022 was obtained from the Bureau of Meteorology. Several parameters like relative humidity (RH), global radiation (Rs), mean, minimum, maximum air temperatures (Ta, Tmin, Tmax), wind speed (WS), and evaporation pan (Ep) were recorded for the three stations. The weather parameters recorded each month concerning quantified weather information gathered from the three stations are specified in Table 1. The table displays Cv, Cx, Xmean, Xmin, Xmax, and Sx, which are the coefficients of variation, skewness, mean, minimum, maximum, and standard deviation, respectively, for the studied meteorological parameters. The table indicates that Ep was the minimum in Brisbane; in contrast, Alice Springs recorded the highest Ep. This pattern could be associated with site differences concerning RH's magnitude, which is inversely associated with evaporation. Brisbane and Alice Springs stations recorded the highest and lowest RH values. In contrast, Adelaide logged maximum skewness. Skewness is positive if the associated data is not proportional and fails to adhere to the standard deviation.

Table 1

Data descriptive analysis

StationDatasetUnitXminXmaxXmeanSxCvCx
Alice Springs Tmax °C 17.36 39.16 29.8 5.97 20 −0.37 
Tmin °C 3.69 23.22 13.8 6.1 44.06 −0.18 
RH 18.2 70.05 42.7 12.65 29.63 0.05 
WS m/s 2.69 4.91 3.7 0.52 13.86 0.11 
Rs MJ m−2 13.8 31.4 22.7 4.84 21.27 −0.09 
Ep mm 2.5 15.09 3.46 38.5 −0.07 
Brisbane Tmax °C 19.9 30.93 25.3 2.95 11.64 −0.09 
Tmin °C 7.68 23.29 15.8 4.38 27.61 −0.09 
RH 57.8 79.96 70.12 4.19 5.97 −0.4 
WS m/s 3.22 7.13 4.5 3.65 80.49 10.22 
Rs MJ m−2 10.13 30.1 19.3 5.02 25.98 0.25 
Ep mm 2.46 9.16 5.5 1.82 32.88 0.16 
Adelaide Tmax °C 14.25 31.86 22.04 4.93 22.37 0.10 
Tmin °C 5.5 19.24 12.18 3.52 28.92 0.11 
RH 46.11 77.01 62.6 8.09 12.91 0.07 
WS m/s 3.25 6.37 4.87 0.59 12.2 −0.16 
Rs MJ m−2 7.08 32.67 18.1 7.14 39.36 0.17 
Ep mm 1.49 10.56 5.33 2.66 49.9 0.25 
StationDatasetUnitXminXmaxXmeanSxCvCx
Alice Springs Tmax °C 17.36 39.16 29.8 5.97 20 −0.37 
Tmin °C 3.69 23.22 13.8 6.1 44.06 −0.18 
RH 18.2 70.05 42.7 12.65 29.63 0.05 
WS m/s 2.69 4.91 3.7 0.52 13.86 0.11 
Rs MJ m−2 13.8 31.4 22.7 4.84 21.27 −0.09 
Ep mm 2.5 15.09 3.46 38.5 −0.07 
Brisbane Tmax °C 19.9 30.93 25.3 2.95 11.64 −0.09 
Tmin °C 7.68 23.29 15.8 4.38 27.61 −0.09 
RH 57.8 79.96 70.12 4.19 5.97 −0.4 
WS m/s 3.22 7.13 4.5 3.65 80.49 10.22 
Rs MJ m−2 10.13 30.1 19.3 5.02 25.98 0.25 
Ep mm 2.46 9.16 5.5 1.82 32.88 0.16 
Adelaide Tmax °C 14.25 31.86 22.04 4.93 22.37 0.10 
Tmin °C 5.5 19.24 12.18 3.52 28.92 0.11 
RH 46.11 77.01 62.6 8.09 12.91 0.07 
WS m/s 3.25 6.37 4.87 0.59 12.2 −0.16 
Rs MJ m−2 7.08 32.67 18.1 7.14 39.36 0.17 
Ep mm 1.49 10.56 5.33 2.66 49.9 0.25 

TNN model input combination scenarios

Choosing the appropriate predictors is critical to devising a robust predictive model (Abed et al. 2023a, 2023b); this study assessed different sets of inputs concerning meteorological variables to successfully develop the suggested TNN input–output model and enhance its predictive characteristics. It is expected to provide a better real-world understanding of input parameters and their impact on the estimated evaporation in the area (Abed et al. 2021a). There are several conscious decisions behind selecting these sets. First, to enable comparison, the TNN model's inputs were selected according to the essential weather factors in the two suggested conventional frameworks (Stephens and Stewart and Thornthwaite). Additionally, predictors were selected after examining the Pearson correlation coefficient (PCC) (Freedman et al. 2007). The PCC technique is a statistical indicator that indicates the mathematical association, or correlation, between two constant variables. It is understood as the most acceptable method to determine correlation interest since it uses the covariance technique (Hauke & Kossowski 2011). It provides information concerning association or correlation magnitude and direction. The two variables may have a negative or positive correlation. The two parameters are not correlated if PCC is 0. More information about the PCC interpretations and ranges can be found in Abed et al. (2022b). This work used PCC to determine the meteorological variables exhibiting the most significant effect on evaporation estimates; the results are shown in Table 2.

Table 2

PCC values for the studied stations

TmaxTminRHWSRsEp
PCC matrix for the Alice Springs station 
Tmax      
Tmin 0.89     
RH −0.55 −0.21    
WS 0.57 0.64 −0.29   
Rs 0.80 0.65 −0.57 0.56  
Ep 0.91 0.77 −0.85 0.67 0.70 
PCC matrix for the Brisbane station 
Tmax      
Tmin 0.94     
RH 0.31 0.54    
WS 0.20 0.19 0.08   
Rs 0.73 0.63 −0.18 −0.05  
Ep 0.82 0.71 −0.86 0.46 0.84 
PCC matrix for the Adelaide station 
Tmax      
Tmin 0.94     
RH −0.89 −0.78    
WS 0.09 0.19 −0.14   
Rs 0.83 0.74 −0.81 0.28  
Ep 0.94 0.88 −0.92 0.39 0.90 
TmaxTminRHWSRsEp
PCC matrix for the Alice Springs station 
Tmax      
Tmin 0.89     
RH −0.55 −0.21    
WS 0.57 0.64 −0.29   
Rs 0.80 0.65 −0.57 0.56  
Ep 0.91 0.77 −0.85 0.67 0.70 
PCC matrix for the Brisbane station 
Tmax      
Tmin 0.94     
RH 0.31 0.54    
WS 0.20 0.19 0.08   
Rs 0.73 0.63 −0.18 −0.05  
Ep 0.82 0.71 −0.86 0.46 0.84 
PCC matrix for the Adelaide station 
Tmax      
Tmin 0.94     
RH −0.89 −0.78    
WS 0.09 0.19 −0.14   
Rs 0.83 0.74 −0.81 0.28  
Ep 0.94 0.88 −0.92 0.39 0.90 

Table 2 illustrates that RH, WS, Rs, Tmax, and Tmin had all been significantly associated with Ep, implying that they might be essential for estimating evaporation parameters for data from each station. Specifically, for every station, Tmax and RH were most profoundly associated with Ep. Hence, Tmax and RH will be considered in all datasets to improve the accuracy of Ep forecasts. Studies have also revealed that Tmax, Tmin, RH, WS, and Rs are the prime factors affecting evaporation (Abed et al. 2022b).

This study examined how the parameter Ep affects evaporation prediction. Data points were chosen based on their correlation with the predicted output. As depicted in Figure 2, the autocorrelation analysis of the monthly Ep time series at each station revealed a significant decrease in correlation beyond the second lag. This suggests that the preceding Ep value two time periods ago notably influenced the predicted Ep value for any given period. Therefore, to enhance the accuracy of the TNN model predictions, the model input included the two most recent pan evaporation records.
Figure 2

Partial autocorrelation (monthly) for the three studied stations (Alice Springs, Brisbane, and Adelaide).

Figure 2

Partial autocorrelation (monthly) for the three studied stations (Alice Springs, Brisbane, and Adelaide).

Close modal

Correspondingly, the present work considered seven distinct input combinations for building the TNN model (Table 3). The climate dataset was partitioned into two parts: 80% was used to train (calibrate) the model, while 20% was used to test (validate). In this regard, to ensure the accuracy of the data partitioning, the TNN models employed the k-fold cross-validation technique. Cross-validation is a method used to assess the performance of AI models by resampling the data into multiple subsets. Each subset serves as a test set while the remaining subsets act as training sets. This approach allows the model to be evaluated on unseen data, which better represents real-world scenarios. The process begins by dividing the data into k equal-sized subsets. During each iteration, k − 1 subsets are used for training the model, and the remaining subset is used for testing. This cycle repeats k times, with each subset being used once as the test set. After completing all iterations, the results from each test are averaged to provide a more accurate estimation of the model's performance. Hence, the dataset was split into two sections based on the initial years, with the first section being used for training and the second section used for validation. This research attempts to conduct a comprehensive assessment to test AI capabilities and employ empirical frameworks to estimate Ep magnitudes at a monthly timescale for three areas across Australia.

Table 3

The input combinations for the TNN model

ModelScenario of inputs
T.N.N-1 Ta 
T.N.N-2 Ta, Rs 
T.N.N-3 RH, Tmax 
T.N.N-4 RH, Tmax, Tmin 
T.N.N-5 Tmax, Tmin, Rs, RH 
T.N.N-6 Tmax, Tmin, Rs, WS, RH 
T.N.N-7 Tmax, Tmin, Rs, WS, RH, Ep 
ModelScenario of inputs
T.N.N-1 Ta 
T.N.N-2 Ta, Rs 
T.N.N-3 RH, Tmax 
T.N.N-4 RH, Tmax, Tmin 
T.N.N-5 Tmax, Tmin, Rs, RH 
T.N.N-6 Tmax, Tmin, Rs, WS, RH 
T.N.N-7 Tmax, Tmin, Rs, WS, RH, Ep 

Empirical models

This study used Stephens and Stewart and Thornthwaite to contrast the two conventional methods. The decision to employ the Stephens and Stewart technique in this research conforms to the recommendations of Sudheer et al. (2002). After thoroughly inspecting 23 widely used techniques for estimating evaporation, they concluded that the Stephens and Stewart model was the most effective of all. Moreover, they are extensively employed approaches (Rosenberry et al. 2004), taking into account the quantity of meteorological inputs and the availability of data.

Stephens and Stewart (SS)

As specified in Equation (1), Stephens & Stewart (1963) recommended more precise outcomes were obtained when measured radiation Rs were used, assuming that data were available. Furthermore, it permits temperature correlations too:
(1)
where Rs, Ep, and Ta represent solar radiation (Cal cm−2 day−1), evaporation (mm), and mean air temperature (F), respectively.

Thornthwaite (TH)

Thornthwaite (1948) conducted an experiment to measure the relationship between evaporation (Ep) and the average monthly temperature (Ta). Uniformity was determined considering 12 sunshine hours per day, while the month comprised 30 days. Probable evaporation (Ep) is computed using the Thornthwaite method; Equation (2) was used to determine the monthly Thornthwaite heat index ():
(2)

where symbolises the median monthly temperature (°C).

The annual heat index is computed by adding the monthly heat indices , as shown in the following equation:
(3)
Equation (4) is used to calculate the potential evaporation Ep for a month:
(4)
where a is computed using the following equation:
(5)
The expression for Ep for a given month is shown in the following equation
(6)

The symbols d and N stand for the number of days in a month and the amount of theoretical sunshine hours, respectively.

Self-attention transformer model for Ep prediction

To address the constraints of recurrent and convolutional sequential methods, the transformer framework incorporates a self-attention mechanism. The transformer architecture utilises self-attention to determine which data are essential to the encoding of the current token by selectively preserving only the most relevant information from the previous token. Otherwise stated, tweaking is conducted for the attention approach for calculating the equivalent of latent space pertaining to the decoder and encoder. Nonetheless, positional encoding needs to be incorporated with the outputs and inputs, a mandate for loss of recurrence. Similarly, position data enables the transformer system to cycle through sequences of inputs and outputs repeatedly in a time-step manner. The encoding layer of the transformer system consists of two components: the feed-forward layer and multi-head self-attention. A one-to-one correlation is established by the attention system, regarding the time-specific moments. The attention layers are stimulated by aspects pertaining to human attention; however, fundamentally, a weighted mean reduction is also included in it. To the attention layer, feeding of three inputs is carried out: query, values, and keys. Each of the sub-layer includes residual associations; consequently, normalisation of the layers occurs. The aim behind various heads is the contrast against employing different Convolutional Neural Network (CNN) filters, in which each filter would extract latent characteristics from the input. Similarly, in the multiheaded attention approach, different latent features pertaining to the input are reduced via various heads. The concatenation function is employed to combine the outputs from every head. The transformer approach, unlike recurrent networks, does not suffer from the vanishing gradient problem and can make reference to an earlier point in the sequence regardless of its range. This characteristic allows the transformer scheme to detect long-term reliance. Furthermore, as opposed to the RNNs, sequential computing is not needed for the transformer, enabling quicker speed employing parallel processing. In other words, the assessment of transformer inputs is not assessed in a sequential manner. Thus, the issue of vanishing gradients is inherently avoided. However, RNNs are not able to perform effectively with regard to long-term forecasts. In comparison, transformers maintain direct associations with each of the previous timesteps, thus permitting data to traverse over an extended series. Nonetheless, a new issue has emerged: there is a direct correlation between the framework with the humongous data inputs. With regard to the transformer framework, the use of a self-attention mechanism is employed to segment non-essential information.

Several research studies have attempted to demonstrate the superior forecasting accuracy of DL frameworks compared with other ML techniques. Thus, the objective of this study is to evaluate the efficacy pertaining to the transformer approach to estimate evaporation with regards to accuracy and efficiency metrics measurement. Nonetheless, the primary transformer model cannot be used as-is to predict time series. To enable the prediction of time series, adjustments have been made to the transformer model. The input now consists of the magnitude of the time series at a given point in time, rather than the embedding layers related to Natural Language Processing (NLP). Additionally, the soft-max classification layer has been eliminated from the output. The output layer has been altered to include a linear activation function, and the mean square error (MSE) has been utilised as the regression-based loss function. In this study, an encoder with two sub-layers is employed for training: self-attention and a one-dimensional convolutional system, which serves as an alternative to the Multi-Layer Perceptron (MLP) layer in the original transformer framework to recognise high-level characteristics. The convolutions do not have dense connections, meaning that not all input nodes impact all output nodes. Additional versatility with regards to learning data attributes is conferred to convolutional layers by such a configuration. In brief, four encoding blocks were connected in a feed-forward manner and comprised the transformer model, as depicted in Figure 3. The encoding system begins with a transformer input time series, which is then processed through a self-attention layer. Following this, layer normalisation is deployed. A feed-forward layer is then utilised in conjunction with a separate normalisation layer process.
Figure 3

TNN model scheme.

Figure 3

TNN model scheme.

Close modal
In this work, an exhaustive search process is utilised for designing the system and conducting hyperparameter training to construct the optimal architecture for the proposed TNN model. Thus, to determine the optimal architecture, several differently configured models were evaluated. The optimal hyperparameters pertaining to the TNN model put forward have been devised by employing four identical transformer encoders. Each of the encoders includes 4 heads of size 64, whose output is processed by employing a 1D global average pooling layer that includes 16 output filters along with a kernel size of 1. Global average pooling is advantageous because it works in tandem with the convolutional architecture, connecting feature maps and classifications. Thus, this can be understood as category confidence indicators. Moreover, the need for parameter optimisation is not required for global average pooling, thereby no overfitting is required (Lin et al. 2013). To the global average pooling, a 128-neuron dense layer is added, which includes the exponential linear unit as an activation function. In addition, to control overfitting, a dropout layer is added with a 0.2 dropout ratio (Ferreira & da Cunha 2020). Finally, employment of one further solitary fully connected linear layer (i.e., forecasting the Ep) is performed. The model was trained in assorted cycles, each with 200 epochs and 16 batches. In order to regulate network weights pertaining to decrease of MSE loss function, the Adam algorithm was employed (Kingma & Ba 2014). Furthermore, at a 1e-2 learning speed, assessment of the network performance was carried out. Post devising the architectures of the framework, training of the system was conducted via validation utilising the information from the training phase. The forecasting ability of the model was assessed relying on the subsequent step by considering the new data. Denormalisation of the forecast numbers was carried out to allow visual representation, this was accompanied by a comparison to the real values. The TNN prediction framework is demonstrated in Figure 4, which outlines the progression of steps involved.
Figure 4

TNN model development procedure.

Figure 4

TNN model development procedure.

Close modal

Performance evaluation

Finding the right performance indicators is essential, as each one possesses its own unique characteristics. Moreover, getting an understanding of the properties regarding every statistical indicator is also helpful to better understand the way a model performs. Thus, in this research, various numerical metrics were used to assess the model's predictive performance, which is described as follows:

  • (1) Coefficient of determination (R2): The R2 denotes how well the estimated and actual outputs match, whose values can range from 0 to 1 (includes both limits). A stochastic framework is denoted by a value of zero, while a perfect fit is signified by a value of one. Since R2 is very popular, comparison of the model becomes easier as well as more consistent. The R2 can be computed using the following equation:
    (7)
  • (2) Root mean square error (RMSE): The average of the squares of the contrast between the real and estimated values, also known as the root mean square error. With respect to the evaluation of performance for the regression model, RMSE has been broadly employed versus MSE. Furthermore, RMSE is known for its simplicity and ease of use for determination. Moreover, big errors are penalised by RMSE, and thus they are regarded to be more acceptable. The RMSE can be computed using the following equation:
    (8)
  • (3) Mean absolute error (MAE): The MAE is calculated by taking the absolute difference between the real and predicted values, as shown in Equation (9). This metric does not penalise large errors that occurred by outliers, as opposed to other metrics such as the mean squared error.
    (9)
  • (4) Nash–Sutcliffe efficiency (NSE): This performance metric is used to compare the amount of residual variance (noise) to the amount of variability in a given set of data. It is often utilised in hydrologic modelling as it can normalise data accuracy, making it easier to understand. The NSE can be computed using the following equation:
    (10)
where n denotes the number of samples, denotes the estimated values, denotes the actual output mean, and denotes the actual output.

Monthly Ep prediction using empirical models

Two conventional models were employed for estimating Ep on a monthly scale, which includes temperature-based as well as radiation-based models. The R2, MAE, RMSE, and NSE values from the two used methods compared with the measured data are shown in Table 4. As signified based on the numerical findings presented in Table 4, the SS (radiation-based) demonstrated greater prediction precision versus the temperature-based model. With the radiation-based model for all stations, the lowest RMSE values were 0.393, 0.398, and 0.383, and the highest R2 values were 0.657, 0.667, and 0.627. With the Thornthwaite model, RMSE values were increased by an average of 21%, while the corresponding R2 values were decreased by an average of 31%. As shown in Table 4, the performance values seen evidently imply that the Thornthwaite model was surpassed by the radiation-based model. This is possible because solar radiation was included, which in general provides an enhancement over those employing only temperature-based estimation (Abed et al. 2021a).

Table 4

The findings of Stephens and Stewart and Thornthwaite models

StationModelR2RMSEMAENSE
Alice Springs Stephens and Stewart 0.657 0.393 0.318 0.676 
Thornthwaite 0.453 0.510 0.410 0.454 
Brisbane Stephens and Stewart 0.667 0.398 0.308 0.668 
Thornthwaite 0.472 0.503 0.408 0.473 
Adelaide Stephens and Stewart 0.627 0.383 0.302 0.628 
Thornthwaite 0.411 0.479 0.373 0.412 
StationModelR2RMSEMAENSE
Alice Springs Stephens and Stewart 0.657 0.393 0.318 0.676 
Thornthwaite 0.453 0.510 0.410 0.454 
Brisbane Stephens and Stewart 0.667 0.398 0.308 0.668 
Thornthwaite 0.472 0.503 0.408 0.473 
Adelaide Stephens and Stewart 0.627 0.383 0.302 0.628 
Thornthwaite 0.411 0.479 0.373 0.412 

Estimation of monthly Ep using the TNN model

To show the robustness of the TNN model pertaining to evaporation prediction, this section presents a complete analysis with regards to the empirical outcomes that have been derived based on the experiment employing this model. In this research, the designed TNN model was used for monthly Ep forecasting at three sites in Australia: Alice Springs, Adelaide, and Brisbane. It is important to highlight that, in agricultural and hydrological contexts, the monthly evaporation rate is essential for calculating water budgets and determining crop water requirements. In Table 5, a significant difference was noted in the accuracy of the Ep forecast depending on the input combinations. The best numerical metrics are highlighted in bold. In fact, the accuracy of a model's prediction could be improved through the use of the full climatological dataset (Ep, RH, Tmin, Tmax, Rs, and WS) from all sites rather than combining inputs with incomplete data. Current findings show that the accuracy of prediction models increases as the number of input variables increases, which corresponds to a previous study by Abed et al. (2022b). To attain good agreement in monthly Ep prediction, four input combinations were adequate. The prediction accuracy of the TNN model was found to be insufficient for all stations when there were only RH and Tmax data used. This demonstrated that employing advanced capabilities like AI may not enhance the model's predictive performance, especially when a limited number of meteorological inputs exist. When five input parameters were employed, adequate results were attained; however, when Ep was used as an input, there was a slight enhancement in the accuracy of the prediction.

Table 5

Evaluation performance of the TNN model for predicting monthly Ep

StationModelR2RMSE (mm)MAE (mm)NSE
Alice Springs T.N.N-1 0.660 0.148 0.122 0.661 
T.N.N-2 0.769 0.121 0.100 0.770 
T.N.N-3 0.830 0.105 0.078 0.831 
T.N.N-4 0.882 0.087 0.067 0.883 
T.N.N-5 0.912 0.075 0.055 0.913 
T.N.N-6 0.935 0.065 0.050 0.936 
T.N.N-7 0.966 0.047 0.039 0.967 
Brisbane T.N.N-1 0.677 0.149 0.125 0.678 
T.N.N-2 0.805 0.116 0.097 0.806 
T.N.N-3 0.847 0.102 0.082 0.848 
T.N.N-4 0.876 0.093 0.075 0.877 
T.N.N-5 0.895 0.088 0.058 0.888 
T.N.N-6 0.934 0.074 0.048 0.925 
T.N.N-7 0.960 0.051 0.039 0.963 
Adelaide T.N.N-1 0.717 0.153 0.130 0.728 
T.N.N-2 0.795 0.133 0.114 0.795 
T.N.N-3 0.874 0.103 0.086 0.875 
T.N.N-4 0.925 0.080 0.065 0.926 
T.N.N-5 0.943 0.070 0.058 0.944 
T.N.N-6 0.954 0.061 0.052 0.955 
T.N.N-7 0.982 0.035 0.027 0.985 
StationModelR2RMSE (mm)MAE (mm)NSE
Alice Springs T.N.N-1 0.660 0.148 0.122 0.661 
T.N.N-2 0.769 0.121 0.100 0.770 
T.N.N-3 0.830 0.105 0.078 0.831 
T.N.N-4 0.882 0.087 0.067 0.883 
T.N.N-5 0.912 0.075 0.055 0.913 
T.N.N-6 0.935 0.065 0.050 0.936 
T.N.N-7 0.966 0.047 0.039 0.967 
Brisbane T.N.N-1 0.677 0.149 0.125 0.678 
T.N.N-2 0.805 0.116 0.097 0.806 
T.N.N-3 0.847 0.102 0.082 0.848 
T.N.N-4 0.876 0.093 0.075 0.877 
T.N.N-5 0.895 0.088 0.058 0.888 
T.N.N-6 0.934 0.074 0.048 0.925 
T.N.N-7 0.960 0.051 0.039 0.963 
Adelaide T.N.N-1 0.717 0.153 0.130 0.728 
T.N.N-2 0.795 0.133 0.114 0.795 
T.N.N-3 0.874 0.103 0.086 0.875 
T.N.N-4 0.925 0.080 0.065 0.926 
T.N.N-5 0.943 0.070 0.058 0.944 
T.N.N-6 0.954 0.061 0.052 0.955 
T.N.N-7 0.982 0.035 0.027 0.985 

The R2 values pertaining to TNN tested at the three studied stations were 0.966 for Alice Springs, 0.982 for Adelaide, and 0.960 for Brisbane. Thus, the TNN model was found to demonstrate the greatest degree of collinearity between observed (Ep) and estimated (Ep) values. Furthermore, R2 values yielded by the transformer model with regard to the three stations were close to 1. It can be noted that the magnitude with regards to the (R2) was 1 when the model delivered ideal performance. Thus, in terms of R2 value, the TNN model demonstrated a superior performance for the prediction Ep values. The findings demonstrate that as the accuracy of the model increased, the MAE and RMSE values decreased. The TNN model displayed improved performance in terms of MAE and RMSE values, with respective values of 0.039, 0.039, 0.027, 0.047, 0.051, and 0.035 mm for all study sites. These metrics measure the accuracy of the model by evaluating the forecasts, and the smaller the values, the superior the performance. Thus, the TNN model was deemed to have achieved superior results. The NSE value is another metric measurement for model accuracy used for assessing the proposed TNN model's efficacy, which seemed to be close to one with regard to all the studied regions. For instance, to focus on this metric concerning the transformer model, NSE ≥0.963 has been observed with regard to all the studied regions. Overall, the current assessment provides convincing evidence confirming the TNN model's significant potential in predicting Ep for all the study locations in Australia.

For all the studied stations in the testing phase, the TNN model demonstrated smaller predictability errors. The scatter plots and time series in Figure 5 compare the forecasted monthly Ep from the transformer model in the testing phase to the observed Ep. The coefficient of determination (R2) of these plots demonstrates how well the observed Ep fits the modelled Ep, with values ranging from 0 to 1. The TNN model had an excellent fit with the observed Ep, with R2 values close to one. However, upon analysing the time series figures, the model demonstrates lower performance, especially in capturing peak values at certain extreme points. Fortunately, these occurrences are limited to a few instances. Overall, the model consistently performs well across all stations, indicating its ability to accurately predict evaporation. Ultimately, as seen in Figure 5, the prediction accuracy of the TNN model was accurate, with R2 values that were very close to 1 for all the studied stations. This deduction showed the high accuracy level shown by the TNN model, in terms of Ep prediction, and it demonstrated the putative DL model to be a key tool for forecasting Ep, showing a positive impact on enhancement of the prediction accuracy due to the self-attention mechanism.
Figure 5

Scatter plot (left) and time series (right) of predicted versus measured Ep for the three studied stations (Alice Springs, Brisbane, and Adelaide).

Figure 5

Scatter plot (left) and time series (right) of predicted versus measured Ep for the three studied stations (Alice Springs, Brisbane, and Adelaide).

Close modal

Comparison of TNN and empirical models

Table 6 shows the outcomes of two empirical models used to forecast monthly Ep, and then compared with their respective TNN models employing the same input combinations. In the initial observation for the input combination of Rs and Ta for all the stations, the lowest prediction accuracy in terms of the R2 values (which were 0.657, 0.667, and 0.627) was offered by the radiation-based model (Stewart and Stephens) versus the TNN model. The performance of the TNN model was seen to be excellent with regards to achieving high prediction accuracy when compared with the Thornthwaite (temperature-based model) when the input was only Ta. The results of Table 6 show that the TNN model exhibits significantly higher performance than the empirical models. It demonstrated a remarkable capability to accurately predict monthly Ep, even when using the same input parameters, due to its ability to handle non-linear and complex functions. This could be due to the TNN self-attention feature that can detect concealed characteristics, which suggests that the transformer architecture is a more powerful approach for evaporation prediction.

Table 6

Evaluation results of the TNN and empirical models

Input combinationStation/ModelR2MAE (mm)RMSE (mm)NSE
 Alice Springs     
Ta, Rs Stephens and Stewart 0.657 0.393 0.318 0.676 
 TNN-2 0.769 0.100 0.121 0.770 
Ta Thornthwaite 0.453 0.510 0.410 0.454 
 TNN-1 0.660 0.122 0.148 0.661 
 Brisbane     
Ta, Rs Stephens and Stewart 0.667 0.398 0.308 0.668 
 TNN-2 0.805 0.097 0.116 0.806 
Ta Thornthwaite 0.472 0.503 0.408 0.473 
 TNN-1 0.677 0.125 0.149 0.678 
 Adelaide     
Ta, Rs Stephens and Stewart 0.627 0.383 0.302 0.628 
 TNN-2 0.795 0.114 0.133 0.795 
Ta Thornthwaite 0.411 0.479 0.373 0.412 
 TNN-1 0.717 0.130 0.153 0.728 
Input combinationStation/ModelR2MAE (mm)RMSE (mm)NSE
 Alice Springs     
Ta, Rs Stephens and Stewart 0.657 0.393 0.318 0.676 
 TNN-2 0.769 0.100 0.121 0.770 
Ta Thornthwaite 0.453 0.510 0.410 0.454 
 TNN-1 0.660 0.122 0.148 0.661 
 Brisbane     
Ta, Rs Stephens and Stewart 0.667 0.398 0.308 0.668 
 TNN-2 0.805 0.097 0.116 0.806 
Ta Thornthwaite 0.472 0.503 0.408 0.473 
 TNN-1 0.677 0.125 0.149 0.678 
 Adelaide     
Ta, Rs Stephens and Stewart 0.627 0.383 0.302 0.628 
 TNN-2 0.795 0.114 0.133 0.795 
Ta Thornthwaite 0.411 0.479 0.373 0.412 
 TNN-1 0.717 0.130 0.153 0.728 

Discussion

Transformers are regarded as a state-of-the-art approach to DL. The advent of the transformer has transformed the application of attention by substituting convolution and recurrence with its unique self-attention mechanism. In terms of time series forecasting, transformers do not analyse their inputs sequentially. Thus, the vanishing gradient issue, which negatively affects the performance of RNNs in long-term prediction, has been addressed by this approach. A new self-attention algorithm has been applied to the evaporative prediction domain in the present research. Good performances were shown by the developed TNN model with respect to prediction Ep applicable to all three chosen stations. The efficacy of the TNN model is evident in its consistency of Ep prediction performance and high accuracy across all the studied sites. This can be attributed to the incorporation of self-attention mechanisms, which are integral to the transformer's architecture. This demonstrates the potential of using intra-attention to enhance the performance of DL predictive models.

As mentioned in the literature, with regards to applicability, the current methods of prediction of open water evaporation differ between and, in certain cases, within regions as well; there is no best method to be generally adopted. In fact, evaporation prediction models face limitations due to variability in environmental factors such as wind speed, humidity, and solar radiation (Shiri & Kişi 2011). They may also struggle with accuracy over diverse geographic regions and lack precision for specific local conditions. In addition, model assumptions and simplifications can also affect reliability, especially in complex terrain or changing climate scenarios (Abed et al. 2022b). As a result, the transformer model, when improved with the self-attention component, is a promising technique for Ep prediction. Additional research can also select a wider range of other regions, with different weather conditions. Nonetheless, DL has far-reaching implications in terms of managing water resource systems and irrigations, as it can be used to monitor monthly variations in evaporation. However, a few challenges need to be considered when developing DL models, including the computational demands, as these models require substantial computational resources, particularly for training large-scale models with extensive datasets. Furthermore, such models require significant amounts of training data during the calibration phase.

The Ep modelling approach, demonstrated by this research, provides a reliable estimate for water losses caused by evaporation, which is essential for managing water resources efficiently. Multiplying Ep values by the land area of watering resources offers an effective scientific-based method to assess the quantity of evaporative water loss, a major factor in determining the existing water asset volume. This calculation simplifies the task of estimating the total amount of existing water available for watering and enables the use of a set of intelligent irrigation schedules. These also help reduce unessential water loss and make the irrigation process more efficient. Thus, this research suggests that the application of the TNN model to predict Ep comes with considerable economic benefits for farmers, especially in areas suffering from drought, water scarcity, or other hydrological imbalances. Additionally, it offers valuable insight for hydrologists about how to incorporate soft computing into their analysis of non-stationary and non-linear hydrological variables.

The objective of this study was to develop a TNN predictive model for monthly evaporation losses and compare it to other empirical models, such as the temperature-based model (Thornthwaite) and the radiation-based model (Stephens and Stewart). The effectiveness of the DL model was assessed by predicting Ep using monthly data from three weather stations located in Australia: Alice Springs, Brisbane, and Adelaide. The monthly Ep from 2009 to 2022 was used as time series data for training (calibration) and testing (validation) of the designed model. The PCC was used to choose the appropriate input parameters (predictors) for the TNN model in terms of Ep forecasting. Conventional evaluation metrics were used to determine the effectiveness of each model.

The investigation led to the following findings:

  • 1. The developed TNN model demonstrated a remarkable level of accuracy when used to forecast monthly Ep values at each of the sites selected in this study.

  • 2. All stations showed the highest prediction accuracy when using models based on comprehensive meteorological datasets, as opposed to models using limited data.

  • 3. The developed TNN model proved to be superior to the empirical methods. Furthermore, when using the same set of inputs as those methods, the accuracy of TNN's monthly Ep projections was significantly improved.

  • 4. To create an Ep prediction model that is highly reliable and widely applicable, it is crucial to evaluate the technique's effectiveness across various regions in Australia and globally. This exploration holds significant promise for future research endeavours. By subjecting the technique to examination in various geographical contexts, the foundation can be laid for a comprehensive understanding of its capabilities and potential, paving the way for robust and versatile advancements in this field of study.

The author would like to thank the Australian Government Research Training Program Scholarship (RTP) for its support.

M.A.: methodology, formal analysis, visualization, and writing – review and editing. M.A.I. and A.N.A.: writing – review and editing and supervision. Y.F.H.: writing – review and editing.

Data cannot be made publicly available; readers should contact the corresponding author for details.

The authors declare there is no conflict.

Abed
M. M.
,
El-Shafie
A.
&
Osman
S. A. B.
2010
Creep predicting model in masonry structure utilizing dynamic neural network
.
Journal of Computer Science
6
(
5
),
597
.
Abed
M.
,
Imteaz
M.
,
Ahmed
A. N.
&
Huang
Y. F.
2021a
Improved prediction of monthly pan evaporation utilising support vector machine technique
. In:
2021 IEEE Asia-Pacific Conference on Computer Science and Data Engineering (CSDE)
, pp.
1
5
.
https://doi.org/10.1109/CSDE53843.2021.9718389
.
Abed
M.
,
Imteaz
M. A.
,
Ahmed
A. N.
&
Huang
Y. F.
2021b
Application of long short-term memory neural network technique for predicting monthly pan evaporation
.
Scientific Reports
11
(
1
),
1
19
.
Abed
M.
,
Imteaz
M.
,
Ahmed
A. N.
&
Huang
Y. F.
2022a
Application of decision tree algorithm for predicting monthly pan evaporation rate
. In
Hydrology & Water Resources Symposium 2022 (HWRS 2022): The Past, the Present, the Future
.
Engineers Australia
,
Brisbane
, pp.
434
445
.
Adnan
R. M.
,
Mostafa
R. R.
,
Dai
H.-L.
,
Heddam
S.
,
Kuriqi
A.
&
Kisi
O.
2022b
Modelling monthly pan evaporation utilising random forest and deep learning algorithms
.
Scientific Reports
12
(
1
),
1
29
.
Adnan
R. M.
,
Mostafa
R. R.
,
Dai
H.-L.
,
Heddam
S.
,
Kuriqi
A.
&
Kisi
O.
2023a
Application of k-nearest neighbors (KNN) technique for predicting monthly pan evaporation
.
AIP Conference Proceedings
2631
(
1
),
020020
.
https://doi.org/10.1063/5.0131894
.
Abed
M.
,
Imteaz
M. A.
&
Ahmed
A. N.
2023b
A comprehensive review of artificial intelligence-based methods for predicting pan evaporation rate
.
Artificial Intelligence Review
56
(
Suppl 2
),
2861
2892
.
Adnan
R.M.
,
Mostafa
R.R.
,
Dai
H.-L.
,
Heddam
S.
,
Kuriqi
A.
&
Kisi
O.
2023
Pan evaporation estimation by relevance vector machine tuned with new metaheuristic algorithms using limited climatic data
.
Engineering Applications of Computational Fluid Mechanics
17
(
1
),
2192258
.
Al-Mukhtar
M.
2021
Modeling of pan evaporation based on the development of machine learning methods
.
Theoretical and Applied Climatology
146
(
3–4
),
961
979
.
https://doi.org/10.1007/s00704-021-03760-4
.
Arif
Z. H.
,
Mahmoud
M. A.
,
Abdulkareem
K. H.
,
Kadry
S.
,
Mohammed
M. A.
,
Al Mhiqani
M. N.
,
Al Waisy
A. S.
&
Nedoma
J.
2022
Adaptive deep learning detection model for multi-foggy images
.
International Journal of Interactive Multimedia & Artificial Intelligence
7
(
7
),
22
37
.
Bengio
Y.
,
Simard
P.
&
Frasconi
P.
1994
Learning long-term dependencies with gradient descent is difficult
.
IEEE Transactions on Neural Networks
5
(
2
),
157
166
.
https://doi.org/10.1109/72.279181
.
Bontempo Scavo
F.
,
Tina
G. M.
,
Gagliano
A.
&
Nižetić
S.
2021
An assessment study of evaporation rate models on a water basin with floating photovoltaic plants
.
International Journal of Energy Research
45
(
1
),
167
188
.
Chang
F. J.
,
Chang
L. C.
,
Huang
C. W.
&
Kao
I. F.
2016
Prediction of monthly regional groundwater levels through hybrid soft-computing techniques
.
Journal of Hydrology
541
,
965
976
.
https://doi.org/10.1016/j.jhydrol.2016.08.006
.
Craig
I.
,
Green
A.
,
Scobie
M.
&
Schmidt
E.
2005
Controlling evaporation loss from water storages. National Centre for Engineering in Agriculture, Toowoomba
.
Ditthakit
P.
,
Pinthong
S.
,
Salaeh
N.
,
Weekaew
J.
,
Tran
T. T.
&
Pham
Q. B
2022
Comparative study of machine learning methods and GR2M model for monthly runoff prediction
.
Ain Shams Engineering Journal
14
(
4
),
101941
.
Falamarzi
Y.
,
Palizdan
N.
,
Huang
Y. F.
&
Lee
T. S.
2014
Estimating evapotranspiration from temperature and wind speed data using artificial and wavelet neural networks (WNNs)
.
Agricultural Water Management
140
,
26
36
.
Ferreira
L. B.
&
da Cunha
F. F.
2020
Multi-step ahead forecasting of daily reference evapotranspiration using deep learning
.
Computers and Electronics in Agriculture
178
,
105728
.
https://doi.org/10.1016/j.compag.2020.105728
.
Freedman
D.
,
Pisani
R.
&
Purves
R.
2007
Statistics
.
WW Norton & Company
,
New York
.
Hauke
J.
&
Kossowski
T.
2011
Comparison of Values of Pearson's and Spearman's Correlation Coefficient on the Same Sets of Data
.
Haykin
S.
&
Lippmann
R.
1994
Neural networks, a comprehensive foundation
.
International Journal of Neural Systems
5
(
4
),
363
364
.
Hu
C.
,
Wu
Q.
,
Li
H.
,
Jian
S.
,
Li
N.
&
Lou
Z.
2018
Deep learning with a long short-term memory networks approach for rainfall-runoff simulation
.
Water
.
https://doi.org/10.3390/w10111543
.
Ikram
R. M. A.
,
Mostafa
R. R.
,
Chen
Z.
,
Islam
A. R. M. T.
,
Kisi
O.
,
Kuriqi
A.
&
Zounemat-Kermani
M.
2022
Advanced hybrid metaheuristic machine learning models application for reference crop evapotranspiration prediction
.
Agronomy
13
(
1
),
98
.
Jasmine
M.
,
Mohammadian
A.
&
Bonakdari
H.
2022
On the prediction of evaporation in arid climate using machine learning model
.
Mathematical and Computational Applications
27
(
2
),
32
.
https://doi.org/10.3390/MCA27020032
.
Keskin
M. E.
&
Terzi
Ö.
2006
Artificial neural network models of daily pan evaporation
.
Journal of Hydrologic Engineering
11
(
1
),
65
70
.
Kingma
D. P.
&
Ba
J.
2014
Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 [Preprint]
.
Kisi
O.
,
Genc
O.
,
Dinc
S.
&
Zounemat-Kermani
M.
2016
Daily pan evaporation modeling using chi-squared automatic interaction detector, neural networks, classification and regression tree
.
Computers and Electronics in Agriculture
122
,
112
117
.
Kisi
O.
,
Mirboluki
A.
,
Naganna
S. R.
,
Malik
A.
,
Kuriqi
A.
&
Mehraein
M.
2022
Comparative evaluation of deep learning and machine learning in modelling pan evaporation using limited inputs
.
Hydrological Sciences Journal
67
(
9
),
1309
1327
.
Kumar
M.
,
Kumari
A.
,
Kumar
D.
,
Al-Ansari
N.
,
Ali
R.
,
Kumar
R.
,
Kumar
A.
,
Elbeltagi
A.
&
Kuriqi
A.
2021
The superiority of data-driven techniques for estimation of daily pan evaporation
.
Atmosphere
12
(
6
),
701
.
https://doi.org/10.3390/ATMOS12060701
.
Lin
M.
,
Chen
Q.
&
Yan
S.
2013
Network in network. arXiv preprint arXiv:1312.4400 [Preprint]
.
McJannet
D. L.
,
Webster
I. T.
,
Stenson
M. P.
&
Sherman
B. S.
2008
Estimating Open Water Evaporation for the Murray-Darling Basin: A Report to the Australian Government From the CSIRO Murray-Darling Basin Sustainable Yields Project
.
CSIRO
,
Melbourne
.
Shao
T.
,
Guo
Y.
,
Chen
H.
&
Hao
Z.
2019
Transformer-based neural network for answer selection in question answering
.
IEEE Access
7
,
26146
26156
.
https://doi.org/10.1109/ACCESS.2019.2900753
.
Stephens
J. C.
&
Stewart
E. H.
1963
A comparison of procedures for computing evaporation and evapotranspiration
.
Publication
62
,
123
133
.
Sudheer
K. P.
,
Gosain
A. K.
,
Mohana Rangan
D.
&
Saheb
S. M.
2002
Modelling evaporation using an artificial neural network algorithm
.
Hydrological Processes
16
(
16
),
3189
3202
.
Thornthwaite
C. W.
1948
An approach toward a rational classification of climate
.
Geographical Review
38
(
1
),
55
94
.
Vaswani
A.
,
Shazeer
N.
,
Parmar
N.
,
Uszkoreit
J.
,
Jones
L.
,
Gomez
A.N.
,
Kaiser
L.
&
Polosukhin
I.
2017
Attention is all you need
.
Advances in Neural Information Processing Systems
.
arXiv preprint. arXiv:1706.03762,10,p.S0140525X16001837
.
Wang
L.
,
Kisi
O.
,
Zounemat-Kermani
M.
&
Gan
Y.
2016
Comparison of six different soft computing methods in modeling evaporation in different climates
.
Hydrology and Earth System Sciences Discussions
1
51
.
Wang
K.
,
He
B.
&
Zhu
W. P.
2021
TSTNN: Two-stage transformer based neural network for speech enhancement in the time domain
. In:
ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing – Proceedings, 2021 June
, pp.
7098
7102
.
https://doi.org/10.1109/ICASSP39728.2021.9413740
.
Wu
L.
,
Huang
G.
,
Fan
J.
,
Ma
X.
,
Zhou
H.
&
Zeng
W.
2020a
Hybrid extreme learning machine with meta-heuristic algorithms for monthly pan evaporation prediction
.
Computers and Electronics in Agriculture
168
,
105115
.
Wu
N.
,
Green
B.
,
Ben
X.
&
O'Banion
S.
2020b
Deep transformer models for time series forecasting: The influenza prevalence case. arXiv preprint arXiv:2001.08317 [Preprint]
.
This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (CC BY-NC-ND 4.0), which permits copying and redistribution for non-commercial purposes with no derivatives, provided the original work is properly cited (http://creativecommons.org/licenses/by-nc-nd/4.0/).