Weirs are designed to stabilize rivers, grade control, and raise upstream water levels. The failure of these structures is primarily due to local scour at the structural site. Consequently, an accurate estimate of the likely scour depth at the structure is critical for weir design safety and economy. This study proposes machine learning models for scour depth prediction at submerged weirs by introducing advanced gradient boosting algorithms, namely gradient boosting (GB), categorical boosting (CatBoost), adaptive boosting (AdaBoost), and extreme gradient boosting (XGBoost). A database consisting of 308 cases was collected for model calibration and evaluation. The results demonstrate that the GB algorithm is very accurate, with coefficients of determination of 0.99610 and 0.96222 for the training and testing datasets, respectively. The GB model outperforms other developed models, such as support vector regression, decision tree, and ridge models, in the literature. A sensitivity analysis study has determined that the morphological jump parameter is the most significant factor, whereas the normal flow depth on the equilibrium bed slope is the least significant factor in predicting the ds under the submerged weir.

  • Proposes machine learning (ML) models, namely gradient boosting (GB), categorical boosting (CatBoost), adaptive boosting (AdaBoost), and extreme gradient boosting (XGBoost), for scour depth prediction at submerged weir.

  • GB model demonstrates superior performance for scour depth prediction at submerged weir.

  • Rigorous statistical metrics validate the robustness of ML-based predictions.

Weirs, also known as bed sills, are river training structures that are used to raise upstream water level, stabilize the bed, and reduce flow velocity. The weir is completely submerged in the river during high-flow events, and scouring occurs both upstream and downstream of the weir. The erosive process during weir overflow may significantly increase local scour downstream and cause severe damage to these structures (Guan 2015). When designing hydraulic structures, it is essential to accurately ascertain the depth of scouring in order to ensure the proper hydraulic design of structures. Previous research has demonstrated that scouring is caused mostly by the materials used to construct river beds, the design of these beds, the timing and geometry of the pier, and the fluid characteristics. As a result, individual characteristics have different effects on scour depth (Najafzadeh et al. 2016a). Various empirical formulations for scouring depth predictions employing effective parameters such as velocity, sediment size, Froude number, and specific sediment weight have been proposed (Vanoni 2006). While empirical formulas are simple to employ, they might either underestimate or overestimate scouring depth. Due to the complexity of the scouring process, empirical formulations cannot be used to obtain a physical understanding of the process. Although numerous empirical formulations for scouring process modeling have been presented, none have achieved sufficient performance due to the interaction of non-linearity, non-stationarity, and stochasticity (Parsaie et al. 2019). As a result, a reliable and accurate physical model capable of simulating the mechanism of the scouring phenomenon has yet to be developed. As a result of these limitations, as well as the need to improve prediction capabilities, various studies have examined novel approaches for improving conventional analysis based on physical properties (Hong et al. 2012; Wei et al. 2022; Wei et al. 2023; Li et al. 2024).

Several laboratory studies were carried out in order to assess the scour depth downstream of grade-control structures. Bormann & Julien (1991) investigated the scouring depth downstream of grade-control structures using large-scale fume tests. D'Agostino & Ferro (2004) analyzed the relationship between upstream head, weir height, and scouring depth.

Artificial intelligence (AI) models have demonstrated their potential in multidisciplinary engineering applications (Li et al. 2021; Luo et al. 2022; Meng et al. 2023; Shi et al. 2023; Zhang et al. 2024). Advances in AI models in the hydraulic engineering discipline have enabled researchers to improve their accuracy when simulating specific phenomena. AI techniques have been used in hydraulic engineering to accurately estimate the scouring process of downstream or adjacent hydraulic structures such as bridge piers, ski-jump buckets, and abutments (Sharafati et al. 2021). Scouring depth prediction is a critical issue for preserving the hydraulic structures. In the literature, different efforts have been initiated for approaching the problem. For example, Sharafati et al. (2021) presented a comprehensive review of the conducted ML methods for the equilibrium scour depth prediction. Where artificial neural networks (ANN), support vector regression (SVR), and M5 tree models were the most used methods. Sharafati et al. (2020) used fuzzy neural networks and particle swarm optimization algorithms for the prediction of scour depth downstream of a sluice gate, which achieved 94% of R2 and 44% of the root mean squared error (RMSE). Furthermore, Parsaie et al. (2019) adopted the SVR algorithm for the prediction of scour depth underneath the river pipelines, which achieved good results in comparison with the ANN and adaptive neuro-fuzzy inference system (ANFIS) algorithms. Nonetheless, Rashki Ghaleh Nou et al. (2019) proposed a self-adaptive extreme learning machine for the scouring depth prediction around the submerged weirs. In which, the proposed approach showed more reliable performance results in comparison with SVR and ANN. Muzzammil et al. (2015) used gene expression programming (GEP) to calculate the scour depth at bridge piers in cohesive, soil showing improvement in prediction performance when compared to Chaudhari's empirical formulations. Najafzadeh et al. (2016a, 2016b) estimated the maximal scouring depth around bridge piers using GEP, model tree, and evolutionary polynomial regression (EPR), taking debris flow into account. According to their report, the EPR models demonstrated superior accuracy compared to the other models. Tao et al. (2021) The author put forth a theoretical framework that combines the extreme gradient boosting model with a genetic algorithm (XGBoost-GA) optimizer. The genetic algorithm (GA) is employed in a hybrid approach to address the hyperparameter optimization challenge in the XGBoost model, in addition to identifying the influential input predictors on ds. The XGBoost–GA model has been developed by incorporating fifteen physical parameters of submerged weirs. The XGBoost–GA model had an excellent degree of predictability, as evidenced by its maximum coefficient of determination (R2 = 0.933) and minimum root mean square error (RMSE = 0.014 m). This field is still being researched and investigated. The innovations and main contributions of this paper are as follows:

  • 1. Providing an accurate and efficient gradient boosting (GB) model for predicting scour depth (ds) at submerged weirs.

  • 2. Examining the prediction accuracy of the best GB model against that of existing machine learning (ML) models, such as SVR, linear regression (LR), decision tree (DT), and ridge models in the literature using performance metrics.

  • 3. To check the correlations among the parameters of ds and to conduct sensitivity analysis to assess the influence of each input parameter on the ds at submerged weirs.

Scour around the submerged weirs

Gaudio et al. (2000) conducted one of the earliest studies on the scouring problem downstream of consecutive bed sills in sloping streams that is based on the assumption of the variability to predict the scour depth (ds) is expressed as follows (Gaudio et al. 2000):
(1)
where ds (m) is the scouring depth downstream of the weir, q (m2/s) is flow rate per unit width, qs (m2/s) is initial volumetric sediment discharge per unit width, ρ (kg/m3) is water density, ρs (kg/m3) is sediment density, hu (m) is normal flow depth on the equilibrium bed slope, v (m/s) is mean velocity component on the transverse direction, g (m/s2) is the acceleration of gravity, d50 (mm) is sediment particle size, a1 (m) is the morphological jump and is defined as:
(2)
where S0 is the initial bed slope (flume slope), Seq is the slope at the equilibrium stage with a grade control structure, and L (m) is the distance between sequent weirs. Lenzi et al. (2002) used the dataset established by Gaudio et al. (2000), and they established the following formulations:
(3)
where Hs is the critical specific energy and Δ is the relative submerged particle density. The SI is calculated based on the sediment size grading:
(4)
Figure 1 displays the definition of the scouring problem for a downstream of consecutive bed sills in a sloping stream (Gaudio et al. 2000).
Figure 1

Scour model sketch for a sequence of weirs in mountain streams with steep slopes.

Figure 1

Scour model sketch for a sequence of weirs in mountain streams with steep slopes.

Close modal
The empirical formulations that have been established were derived by using a wide range of physical data. However, this research development considers the input factors based on the availability and consistency of the acquired data from the literature. Scour depth (ds, m) is simulated based on channel flow rate (Q, m3/s) flow rate per unit width (q, m2/s), normal flow depth on the equilibrium bed slope (hu, m), sediment particle size (d50, mm; d16, mm, and d90,, mm), the morphological jump (a1, m), the initial bed slope (flume slope) (S0), the slope at equilibrium stage with grade control structure (Seq), the distance between sequent weirs (L, m), sorting index (SI), weir width (b, m), channel width (B, m), sediment density (ρs, kg/m3), and the critical specific energy (Hs, m) and expressed as follows (Tao et al. 2021):
(5)
The empirical formulas presented in the literature for depth scouring modeling are associated with various limitations. One of the most common is the assumption that they are suitable for all forms of scouring phenomena. This appears to be due to differences in physical properties. As a result, it is highly recommended to explore more reliable and resilient intelligent models capable of mimicking the actual mechanism of the scouring pattern.

Data collection and correlation analysis

A total of 308 data are used for developing the proposed models (see Supplementary file, Table S1). Data were taken from previously published studies and reflected in Guan (2015). Readers may refer to Guan (2015) for further details. Table 1 presents the mean, standard deviation, kurtosis, skewness, minimum, maximum, and count of the input and output parameters. A lower standard deviation value, such as Q, indicates that the findings are mostly close to the mean, whereas a higher standard deviation value, such as d90, shows that the results are more spread out (Sun et al. 2015; Edjabou et al. 2017; Sun et al. 2018a; Jiang et al. 2024; Liu et al. 2024; Zi et al. 2024). Skewness, which can be positive, zero, negative, or undefined, aids in determining the degree of asymmetry of a probability distribution in the case of a real-valued arbitrary parameter from the perspective of its average value (Sun et al. 2018b; Sharma & Ojha 2020; Rong et al. 2022). Additionally, according to Brown and Greene, kurtosis is usually between −10 (heavy-tailed) and +10 (light-tailed), which helps determine the form of a probability distribution (Brown & Greene 2006; Wang et al. 2017; Zhou et al. 2022; Wang et al. 2024). The kurtosis values for q and a1 are negative and range between −0.6 and −0.2 (following mesokurtic distribution), whereas the rest are positive values (following leptokurtic distribution) (Benson 1993; Lee & Ahn 2019). The correlations among different parameters in the dataset are illustrated in Table 2. The Pearson correlation coefficient (r) was used to correlate all of the investigated input and target variables. Each cell in the plot contains a correlation coefficient, quantifying the degree of association between two parameters. These coefficients range from −1 to 1, indicating the strength and direction of the correlation. From Table 2, it is clear that the parameter a1 (r = 0.854) has a very strong positive correlation and parameter hu (r = −0.004) has a very weak negative correlation with ds. In the order of strong positive to very weak negative, the relationships between input and output parameters are shown in Table 2. As a result, no parameters from the scour depth prediction under the submerged weir were removed.

Table 1

Dataset parameter statistics

QqBbhuS0Seqa1HsLd16d50d90SIρsds
Mean 0.02249 0.03892 0.61968 0.61968 0.02265 0.05630 0.02320 0.04165 0.08599 1.71766 2.19605 7.81981 14.63182 1.81146 2649.22078 0.13495 
Standard deviation 0.01217 0.01203 0.47146 0.47146 0.02709 0.02836 0.01874 0.02414 0.02329 1.07764 0.88231 2.15049 8.53155 1.24110 3.87634 0.06857 
Kurtosis 8.80083 −0.51114 11.01298 11.01298 9.70729 1.34812 1.51905 −0.26140 2.81797 6.73825 14.40440 3.55293 4.20653 6.95659 21.06707 −1.00049 
Skewness 2.56582 −0.32630 3.55377 3.55377 3.28167 0.42015 0.73911 0.64477 0.77819 2.15831 3.39181 −2.31181 1.89967 2.97103 −4.78856 0.42547 
Minimum 0.004 0.007 0.3 0.3 0.013 0.006 0.004 0.025 0.5 1.8 2.3 1.15 2630 0.024 
Maximum 0.081 0.061 2.44 2.44 0.14 0.148 0.104 0.115 0.175 6.5 6.3 8.7 40 5.88 2650 0.298 
Count 308 308 308 308 308 308 308 308 308 308 308 308 308 308 308 308 
QqBbhuS0Seqa1HsLd16d50d90SIρsds
Mean 0.02249 0.03892 0.61968 0.61968 0.02265 0.05630 0.02320 0.04165 0.08599 1.71766 2.19605 7.81981 14.63182 1.81146 2649.22078 0.13495 
Standard deviation 0.01217 0.01203 0.47146 0.47146 0.02709 0.02836 0.01874 0.02414 0.02329 1.07764 0.88231 2.15049 8.53155 1.24110 3.87634 0.06857 
Kurtosis 8.80083 −0.51114 11.01298 11.01298 9.70729 1.34812 1.51905 −0.26140 2.81797 6.73825 14.40440 3.55293 4.20653 6.95659 21.06707 −1.00049 
Skewness 2.56582 −0.32630 3.55377 3.55377 3.28167 0.42015 0.73911 0.64477 0.77819 2.15831 3.39181 −2.31181 1.89967 2.97103 −4.78856 0.42547 
Minimum 0.004 0.007 0.3 0.3 0.013 0.006 0.004 0.025 0.5 1.8 2.3 1.15 2630 0.024 
Maximum 0.081 0.061 2.44 2.44 0.14 0.148 0.104 0.115 0.175 6.5 6.3 8.7 40 5.88 2650 0.298 
Count 308 308 308 308 308 308 308 308 308 308 308 308 308 308 308 308 
Table 2

Correlation among different parameters in the dataset

QqBbhuS0Seqa1HsLd16d50d90SIρsds
Q                
q 0.230               
B 0.844 −0.292              
b 0.844 −0.292 1.000             
hu 0.741 −0.328 0.906 0.906            
S0 −0.315 0.020 −0.329 −0.329 −0.482           
Seq −0.327 −0.196 −0.224 −0.224 −0.308 0.464          
a1 0.132 0.297 −0.012 −0.012 −0.118 0.534 0.037         
Hs 0.936 0.515 0.647 0.647 0.585 −0.352 −0.364 0.180        
L 0.520 −0.229 0.668 0.668 0.715 −0.586 −0.134 0.024 0.4458       
d16 0.816 0.048 0.758 0.758 0.657 −0.376 −0.232 0.027 0.7564 0.465      
d50 0.164 0.442 −0.087 −0.087 −0.346 0.588 0.390 0.362 0.2377 −0.409 0.210     
d90 −0.171 −0.182 −0.090 −0.090 −0.227 0.778 0.475 0.186 −0.3032 −0.432 −0.320 0.447    
SI −0.225 −0.389 −0.029 −0.029 −0.084 0.622 0.366 0.063 −0.4094 −0.295 −0.413 0.098 0.933   
ρs 0.128 0.235 0.008 0.008 0.067 0.342 0.224 0.219 0.2301 −0.146 0.159 0.565 0.292 0.098  
ds 0.261 0.428 0.058 0.058 −0.004 0.208 0.037 0.854 0.3684 0.251 0.133 0.249 −0.124 −0.235 0.150 
QqBbhuS0Seqa1HsLd16d50d90SIρsds
Q                
q 0.230               
B 0.844 −0.292              
b 0.844 −0.292 1.000             
hu 0.741 −0.328 0.906 0.906            
S0 −0.315 0.020 −0.329 −0.329 −0.482           
Seq −0.327 −0.196 −0.224 −0.224 −0.308 0.464          
a1 0.132 0.297 −0.012 −0.012 −0.118 0.534 0.037         
Hs 0.936 0.515 0.647 0.647 0.585 −0.352 −0.364 0.180        
L 0.520 −0.229 0.668 0.668 0.715 −0.586 −0.134 0.024 0.4458       
d16 0.816 0.048 0.758 0.758 0.657 −0.376 −0.232 0.027 0.7564 0.465      
d50 0.164 0.442 −0.087 −0.087 −0.346 0.588 0.390 0.362 0.2377 −0.409 0.210     
d90 −0.171 −0.182 −0.090 −0.090 −0.227 0.778 0.475 0.186 −0.3032 −0.432 −0.320 0.447    
SI −0.225 −0.389 −0.029 −0.029 −0.084 0.622 0.366 0.063 −0.4094 −0.295 −0.413 0.098 0.933   
ρs 0.128 0.235 0.008 0.008 0.067 0.342 0.224 0.219 0.2301 −0.146 0.159 0.565 0.292 0.098  
ds 0.261 0.428 0.058 0.058 −0.004 0.208 0.037 0.854 0.3684 0.251 0.133 0.249 −0.124 −0.235 0.150 

Methodology

Extreme gradient boosting

XGBoost is a method developed by Chen that is based on GB (Chen & Guestrin 2016). The decision trees classifier is typically employed as a weak model in this technique (Zounemat-Kermani et al. 2021). The predictions are derived from a series of weak learners that continuously enhance the performance of their predecessors. XGBoost incorporates a regularization term into the objective function as a means of mitigating the issue of overfitting (Chandrahas et al. 2022).
(6)
where O is the objective function, R(fk) denotes the regularization term at the k iteration time, and C is a constant. The regularization term is expressed as (Chandrahas et al. 2022):
(7)
where α denotes the complexity of leaves, H represents the number of leaves, η denotes the penalty parameter, and wj is the output of each leaf node. The XGBoost algorithm partitions the trees either based on their depth or level-wise. Every tree inside the decision-making process evaluates the feature and its corresponding threshold, as well as determining the most optimal branch effect. The tree topologies are expanded through the use of consecutive splits.

Categorical boosting

CatBoost is a permutation-based alternative to conventional algorithms. It is an innovative approach to handling categorical information in data processing (Dorogush et al. 2018; Prokhorenkova et al. 2018). The suggested method introduces two novel concepts, namely ordered target statistics and ordered boosting. Hancock & Khoshgoftaar (2020) have published an in-depth investigation of this technique, examining its applicability in classification and regression tasks across several domains. CatBoost employs target statistics as additional numerical features to handle category features, a technique that has been demonstrated to be highly effective while minimizing the loss of information (Prokhorenkova et al. 2018). The algorithm produces a randomized arrangement of the dataset, followed by the computation of the mean label value for the training samples belonging to the same category inside the arrangement. Following Prokhorenkova et al. (2018), if σ = (σ1, σ2, … , σn) is a permutation, the category can be substituted with the average label value (Prokhorenkova et al. 2018):
(8)
where P is a prior value; a is the weight of the prior; Yσj is a label value; [·] denotes the Iverson bracket, namely [xσj,kxσp,k] equals 1 if x σj,k = x σp,k, and otherwise, it is equal to 0. For further details regarding CatBoost, interested readers are referred to the publications of Prokhorenkova et al. (2018) and Dorogush et al. (2018).

Adaptive boosting

The ensemble method known as AdaBoost, which combines many weak learner decision trees, demonstrates slight advantages over random guessing. The AdaBoost methodology exhibits adaptability by leveraging the gradient information from previous trees to subsequent trees, with the objective of reducing the error of the preceding tree. AdaBoost exhibits increased resilience against outliers and irrelevant data because of its significant versatility. Moreover, the methodology is specifically devised to function in a manner whereby the following trees are furnished with the information obtained by preceding trees. This allows individuals to focus just on training samples that present challenges in terms of prediction (Freund & Schapire 1997). Figure 2 describes the overall calculation process of the AdaBoost algorithm. When training each basic tree model, the weight distribution of each sample in the dataset needs to be adjusted. Since each training data will change, the training results will also be different, and finally, all of the results are summed (Schapire 2013).
(9)
where Fn(x) is the overall model, Fm−1(x) is the overall obtained in the previous round, yi is the prediction result of the ith tree, and h(xi) is the newly added tree.
Figure 2

AdaBoost algorithm calculation process.

Figure 2

AdaBoost algorithm calculation process.

Close modal

Gradient boosting

GB is a type of ensemble method in which multiple weak models are developed and then combined to improve overall performance. GB uses the methodology of gradient descent in order to minimize the loss function that relates to a given model. The process of incorporating weak learners into the model is carried out using an iterative approach. The ultimate prediction is established by the combined input of every weak learner, which is then determined by a gradient optimization procedure with the objective of reducing the overall error of the strong learner (Aurélien 2019; Islam & Amin 2020). The gradient component is an essential part of gradient boosters. Instead of employing the parameters of the weaker models, the gradient descent optimization approach is utilized on the output of the model. The GB approach is an improved version of the gradient descent technique that enables generalization through the modification of both the gradient and the loss function (Ngo et al. 2023).

Performance evaluation

The evaluation stage involves the computation of diverse assessment metrics, encompassing mean squared error, RMSE, mean absolute error (MAE), coefficient of determination (R2), mean absolute percentage error (MAPE), mean absolute deviation (MAD), and Kling–Gupta efficiency (KGE). These statistical metrics serve to gauge the efficacy of the model's performance, shedding light on the extent to which the model's predictions correlate with the actual target values. The formulations used to calculate these performance metrics are expressed in the following equations (Daneshfaraz et al. 2021a, 2021b; Ahmad et al. 2023a; Abdullah et al. 2024; Al-Zubi et al. 2024; Kalateh et al. 2024a, 2024b):
(10)
(11)
(12)
(13)
(14)
(15)
where di is the ith observed value, yi is the ith predicted value, and dmean is the mean value of the observed values. In Equation (14), R is the linear correlation between observations and predictions, β is a measure of the flow variability error, and γ is a bias term. KGE = 1 indicates perfect agreement between predictions and observations. Figure 3 depicts the proposed methodology adopted in this study.
Figure 3

Flowchart of the proposed methodology.

Figure 3

Flowchart of the proposed methodology.

Close modal

The proposed models that predict the scour depth are developed using orange software. The predictor variables were provided via an input set (x) defined by x = [Q, q, hu, d50, d16, d90, a1, S0, Seq, L, SI, b, B, ρs, and Hs], while the target variable (y) is scour depth (ds) at a submerged weir. Every modeling stage requires the selection of a suitable size of training and testing datasets. Consequently, 80% (246 cases) of the total data were employed to generate models, while the remaining 20% (62 cases) of the data were used to test the developed models in this study. The proposed models were tuned through trial and error to get optimal hyperparameter values owing to accurate prediction of scour depth (ds) at the submerged weir. This study optimizes some essential model parameters and clarifies the definitions of these hyperparameters. The tuning parameters for the models were selected and then changed during the trials until the best metrics presented in Table 3 were obtained.

Table 3

Hyperparameter optimization results

ModelHyperparameter optimization value
XGBoost Number of trees = 20, Learning rate = 0.5, Regularization (λ) = 0.011
Limit depth of individual trees = 10 
GB Number of trees = 110, Learning rate = 0.1
Limit depth of individual trees = 9, Do not split subsets smaller than 5 
AdaBoost Number of estimators = 95, Learning rate = 1, Classification algorithm = SAMME.R, Regression loss function = Linear 
CatBoost Number of trees = 30, Learning rate = 0.5, Regularization (λ) = 3
Limit depth of individual trees = 10 
ModelHyperparameter optimization value
XGBoost Number of trees = 20, Learning rate = 0.5, Regularization (λ) = 0.011
Limit depth of individual trees = 10 
GB Number of trees = 110, Learning rate = 0.1
Limit depth of individual trees = 9, Do not split subsets smaller than 5 
AdaBoost Number of estimators = 95, Learning rate = 1, Classification algorithm = SAMME.R, Regression loss function = Linear 
CatBoost Number of trees = 30, Learning rate = 0.5, Regularization (λ) = 3
Limit depth of individual trees = 10 

Comparing the effectiveness of constructed models

Several performance indices, such as the R2, RMSE, MAE, MAPE, and MAD, were used to evaluate the effectiveness of the utilized models. The performance of the developed models is summarized in Table 4 and in Figures 4 and 5 for both the training and testing phases. From the results, it can be concluded that the proposed GB model attained the highest value of the coefficient of determination R2 = 0.99610 and the lowest value of RMSE = 0.00419, followed by the XGBoost (R2 = 0.99605 and RMSE = 0.00422), AdaBoost (R2 = 0.98985 and RMSE = 0.00676) and CatBoost (R2 = 0.98985 and RMSE = 0.00676) models in the training phase. However, in the testing phase, the AdaBoost model achieves R2 = 0.96539 and RMSE = 0.01339, followed by the GB (R2 = 0.96222 and RMSE = 0.01399), XGBoost (R2 = 0.95984 and RMSE = 0.01442), and CatBoost (R2 = 0.94567 and RMSE = 0.01677) models. The scatter plot between the actual scour depth and the predicted scour depth along the line (x = y). A perfect prediction of the model's performance is represented by a point lying on the line (x = y), and a prediction that is closer to the line (x = y) indicates a more accurate model. The coefficient of determination (R2) and the following performance indices are also presented for each model separately for the training and testing data shown in Figures 4 and 5, respectively.
Table 4

Results of rank analysis based on performance indices

XGBoost
GB
AdaBoost
CatBoost
ParameterTRTSTRTSTRTSTRTS
R2 0.99605 0.95984 0.99610 0.96222 0.98956 0.96539 0.98985 0.94567 
Score 
RMSE 0.00422 0.01442 0.00419 0.01399 0.00686 0.01339 0.00676 0.01677 
Score 
MAE 0.00175 0.01123 0.00142 0.01084 0.00405 0.01035 0.00499 0.01201 
Score 
MAPE 1.74427 12.73715 1.42102 12.81763 4.62190 10.82858 4.78914 16.03426 
Score 
MAD 0.00175 0.01123 0.00142 0.01084 0.00405 0.01035 0.00499 0.01201 
Score 
KGE 0.996 0.976 0.997 0.970 0.966 0.964 0.987 0.913 
Score 
Sub total 18 15 24 17 22 
Total score 33 37 31 15 
Rank 
XGBoost
GB
AdaBoost
CatBoost
ParameterTRTSTRTSTRTSTRTS
R2 0.99605 0.95984 0.99610 0.96222 0.98956 0.96539 0.98985 0.94567 
Score 
RMSE 0.00422 0.01442 0.00419 0.01399 0.00686 0.01339 0.00676 0.01677 
Score 
MAE 0.00175 0.01123 0.00142 0.01084 0.00405 0.01035 0.00499 0.01201 
Score 
MAPE 1.74427 12.73715 1.42102 12.81763 4.62190 10.82858 4.78914 16.03426 
Score 
MAD 0.00175 0.01123 0.00142 0.01084 0.00405 0.01035 0.00499 0.01201 
Score 
KGE 0.996 0.976 0.997 0.970 0.966 0.964 0.987 0.913 
Score 
Sub total 18 15 24 17 22 
Total score 33 37 31 15 
Rank 

Note. TR, training; TS, testing.

Figure 4

The predicted versus actual scour depth (ds) (a) XGBoost, (b) GB, (c) AdaBoost, and (d) CatBoost models based on the training dataset.

Figure 4

The predicted versus actual scour depth (ds) (a) XGBoost, (b) GB, (c) AdaBoost, and (d) CatBoost models based on the training dataset.

Close modal
Figure 5

The predicted versus actual scour depth (ds) (a) XGBoost, (b) GB, (c) AdaBoost, and (d) CatBoost models based on the testing dataset.

Figure 5

The predicted versus actual scour depth (ds) (a) XGBoost, (b) GB, (c) AdaBoost, and (d) CatBoost models based on the testing dataset.

Close modal
The accuracy of all developed models at predicting scour depth values is depicted in Figure 6(a)–6(d) for the training dataset and Figure 7(a)–7(d) for the testing dataset. The GB model (see Figure 6(b)) provided the most reliable prediction, as seen in this graphs whereas in the testing dataset, the AdaBoost model (see Figure 7(c)) showed reliable prediction. Generally, except for a few noise points, it is evident by the higher aggregation of results around the y-axis (y = 0) by the GB model in training and testing datasets. In contrast to the other models, namely XGBoost, AdaBoost, and CatBoost, the comparison results are sufficiently consistent, indicating that the proposed GB is capable of predicting scour depth values.
Figure 6

Comparison of the proposed models results in the training dataset: (a) XGBoost, (b) GB, (c) AdaBoost, and (d) CatBoost in predicting scour depth values.

Figure 6

Comparison of the proposed models results in the training dataset: (a) XGBoost, (b) GB, (c) AdaBoost, and (d) CatBoost in predicting scour depth values.

Close modal
Figure 7

Comparison of the proposed models results in the testing dataset: (a) XGBoost, (b) GB, (c) AdaBoost, and (d) CatBoost in predicting scour depth values.

Figure 7

Comparison of the proposed models results in the testing dataset: (a) XGBoost, (b) GB, (c) AdaBoost, and (d) CatBoost in predicting scour depth values.

Close modal

Comparison of the developed models with available ML models in literature

The results of the current research were also validated against literature reports on the implementation of AI models over the test modeling phase. Figure 8 represents the comparative performance of ML algorithms, for instance, the XGBoost–GA, XGBoost-Grid, SVR, LR, DT, and Ridge models were studied to evaluate the suitability of scouring predictions for submerged weirs. According to the results, XGBoost–GA achieved the maximum R2 (0.933), followed by LR (0.900), DT (0.860), XGBoost-Grid (0.692), Ridge (0.450), and SVR (0.239) (Tao et al. 2021). The proposed GB model (R2 = 0.99610 and RMSE = 0.00419) was demonstrated as being a robust model that can be implemented in the future for hydraulic structure engineering design applications.
Figure 8

Comparative performance of developed models with the available models in literature.

Figure 8

Comparative performance of developed models with the available models in literature.

Close modal

Taylor diagram

In this section, we present a Taylor diagram to evaluate the efficacy of the suggested models by comparing the actual scour depth with the predicted scour depth. The Taylor diagram is a mathematical graphic that is utilized to visually represent the RMSE, coefficient of correlation (R), and standard deviation. Its purpose is to offer a concise evaluation of the model's performance by displaying these metrics simultaneously. The degree of accuracy exhibited by a model is directly proportional to its proximity to the ‘observed’ spot. The standard deviation of models that fall within the dashed black arc exhibits a closer alignment with the actual pattern of experimental data. In Figure 9 it can be seen that the XGBoost, GB, AdaBoost, and CatBoost algorithms' standard deviations are closer to the experimental data's standard deviation. However, the CatBoost algorithm has a slightly higher correlation coefficient with experimental data. Although the GB model is the one that most closely matches experimental data (its points are near the point that shows as ‘observed’ on the x-axis).
Figure 9

Taylor diagram for proposed models.

Figure 9

Taylor diagram for proposed models.

Close modal

Rank analysis

The frequently employed approach to assess the performance of models is through rank analysis, which involves evaluating the models based on their performance index values. In the present methodology, a score of ‘n’ (in the context of this specific study, n = 4, representing a number of computational models) is assigned to the proposed computational model that exhibits the highest performance parameter value. Conversely, a score of 1 is assigned to the model that demonstrates the lowest performance parameter value. This scoring system is applied independently to both the training and testing outcomes. In the context of performance metrics such as RMSE, MAE, and MAPE, the model that achieves the lowest value of error parameters for both the training and testing phases is assigned the highest score (n = 4). Next, the individual scores for each model are put together to produce the models' overall score. The final score of the model is computed by summing the scores obtained from both the training (TR) and testing (TS) phases. The models are ranked in descending order based on their total scores, with the model achieving the greatest score being regarded as the most efficient overall. The details of the score analysis are presented in Table 4, where it is evident that the GB model achieved the highest total score (total score = 37), followed by XGBoost (total score = 33), AdaBoost (total score = 31), and CatBoost (total score = 15).

Sensitivity analysis

Sensitivity analysis is a technique that helps to understand how different factors or inputs affect the output of a model, i.e. GB. The Cosine amplitude approach is used to calculate the relationship between the resilient modulus and the input parameters. It is used in numerous research works to find the most sensitive parameter (Ahmad et al. 2021a, 2021b, 2021c; Ahmad et al. 2022; Amjad et al. 2022; Khan et al. 2022; Ahmad et al. 2023a, 2023b; Abdullah et al. 2024; Al-Zubi et al. 2024). The following equation is used to calculate the degree of sensitivity index for each input parameter.
(16)
where shows the degree of sensitivity index of each input parameter, presents the independent variable for the dependent variable. and represents the dependent variable for the data point. Strong correlation between the input and output variable is indicated by an r value close to 1, whereas weak correlation between the input and output variable is indicated by a r value close to 0. The a1 has the highest degree of sensitivity index (0.961), as seen in Figure 10, which means that varying the a1 has the most impact on ds. On the other hand, hu has the lowest degree of sensitivity index (0.573), which means that changing hu has the least impact on ds. The degree of importance can be presented as a1 > q > Hs > ρs > d50 > d16 > Q > S0 > L > d90 > B and b > Seq > SI > hu.
Figure 10

Degree of sensitivity analysis for predicted ds.

Figure 10

Degree of sensitivity analysis for predicted ds.

Close modal

In this research study, ML algorithms such as XGBoost, GB, AdaBoost, and CatBoost were used to predict the ds under the submerged weir. The performance of the developed models was evaluated using statistical metrics such as R2, RMSE, MAE, MAPE, MAD, and KGE and compared to the recently developed models in literature such as XGBoost–GA, LR, and SVR. The following are the main findings based on the results:

  • (1) Pearson correlation coefficient showed that morphological jump (a1) (r = 0.854) has a very strong positive correlation and normal flow depth on the equilibrium bed slope (hu) (r = −0.004) has a very weak negative correlation with scour depth (ds) for the simulated data.

  • (2) The new proposed models, i.e. XGBoost, GB, AdaBoost, and CatBoost have the highest performance capability as compared to available XGBoost–GA, XGBoost-Grid, SVR, LR, DT, and Ridge models developed recently in literature with less variation in the actual and predicted values in terms of errors in the training and test sets.

  • (3) Results showed that all the R2 values of the four ML algorithms (i.e. XGBoost, GB, AdaBoost, and CatBoost) were larger than 0.90 for both the training dataset and testing dataset, indicating that the proposed approach is able to predict the ds under submerged weir with satisfactory accuracy. Furthermore, the GB model performs best with (R2 = 0.99610, RMSE = 0.00419. MAE = 0.00142, MAPE = 1.42102, KGE = 0.997, and MAD = 0.00142) for the training set. Based on the scatter plots of actual and predicted values, the GB model exhibited a better fit to the actual data, indicating that it has potential for broader applications in scour depth prediction.

  • (4) The sensitivity analysis indicated that the a1 has the highest degree of sensitivity index (0.961), whereas the hu has the lowest degree of sensitivity index (0.573), which contributes to the scour depth prediction.

The accuracy and reliability of predictions provided by the presented models improve when interpolation is employed, as opposed to extrapolation, owing to the use of input values. Therefore, the models should not be applied to input parameter values outside of the range specified by the study. It should be noted that the accuracy and reliability of MLalgorithms are affected by the dataset, such as the number and kind of samples. Therefore, additional samples should be collected and more effective models should be suggested in the future.

The authors are thankful to the Deanship of Graduate Studies and Scientific Research at the Najran University for funding this work under the Easy Funding Program grant code (NU/EFP/SERC/13/132).

All relevant data are included in the paper or its Supplementary Information.

The authors declare there is no conflict.

Abdullah
G. M. S.
,
Ahmad
M.
,
Babur
M.
,
Badshah
M. U.
,
Al-Mansob
R. A.
,
Gamil
Y.
&
Fawad
M.
(
2024
)
Boosting-based ensemble machine learning models for predicting unconfined compressive strength of geopolymer stabilized clayey soil
,
Scientific Reports
,
14
,
2323
.
Ahmad
M.
,
Ahmad
F.
,
Wróblewski
P.
,
Al-Mansob
R. A.
,
Olczak
P.
,
Kamiński
P.
,
Safdar
M.
&
Rai
P.
(
2021a
)
Prediction of ultimate bearing capacity of shallow foundations on cohesionless soils: a Gaussian process regression approach
,
Applied Sciences
,
11
,
10317
.
Ahmad
M.
,
Kamiński
P.
,
Olczak
P.
,
Alam
M.
,
Iqbal
M. J.
,
Ahmad
F.
,
Sasui
S.
&
Khan
B. J.
(
2021b
)
Development of prediction models for shear strength of rockfill material using machine learning techniques
,
Applied Sciences
,
11
,
6167
.
Ahmad
M.
,
Hu
J.-L.
,
Ahmad
F.
,
Tang
X.-W.
,
Amjad
M.
,
Iqbal
M. J.
,
Asim
M.
&
Farooq
A.
(
2021c
)
Supervised learning methods for modeling concrete compressive strength prediction at high temperature
,
Materials
,
14
,
1983
.
Ahmad
M.
,
Al-Mansob
R. A.
,
Ramli
A. B. B.
,
Ahmad
F.
&
Khan
B. J.
(
2023a
)
Unconfined compressive strength prediction of stabilized expansive clay soil using machine learning techniques
,
Multiscale and Multidisciplinary Modeling, Experiments and Design
, 7, 217–231.
Ahmad
M.
,
Al-Zubi
M. A.
,
Kubińska-Jabcoń
E.
,
Majdi
A.
,
Al-Mansob
R. A.
,
Sabri
M. M. S.
,
Ali
E.
,
Naji
J. A.
,
Elnaggar
A. Y.
&
Zamin
B.
(
2023b
)
Predicting California bearing ratio of HARHA-treated expansive soils using Gaussian process regression
,
Scientific Reports
,
13
,
13593
.
Al-Zubi
M. A.
,
Ahmad
M.
,
Abdullah
S.
,
Khan
B. J.
,
Qamar
W.
,
Abdullah
G. M. S.
,
González-Lezcano
R. A.
,
Paul
S.
,
El-Gawaad
N. S. A.
&
Ouahbi
T.
(
2024
)
Long short term memory networks for predicting resilient modulus of stabilized base material subject to wet-dry cycles
,
Scientific Reports
,
14
,
27928
.
Amjad
M.
,
Ahmad
I.
,
Ahmad
M.
,
Wróblewski
P.
,
Kamiński
P.
&
Amjad
U.
(
2022
)
Prediction of pile bearing capacity using XGBoost algorithm: modeling and performance evaluation
,
Applied Sciences
,
12
,
2126
.
Aurélien
G.
(
2019
)
Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems
.
O'Reilly Publishing, Boston, United States
.
Benson
C. H.
(
1993
)
Probability distributions for hydraulic conductivity of compacted soil liners
,
Journal of Geotechnical Engineering
,
119
,
471
486
.
Bormann
N. E.
&
Julien
P. Y.
(
1991
)
Scour downstream of grade-control structures
,
Journal of Hydraulic Engineering
,
117
,
579
594
.
Brown
S. C.
&
Greene
J. A.
(
2006
)
The wisdom development scale: translating the conceptual to the concrete
,
Journal of College Student Development
,
47
,
1
19
.
Chandrahas
N. S.
,
Choudhary
B. S.
,
Teja
M. V.
,
Venkataramayya
M. S.
&
Prasad
N. K.
(
2022
)
XG boost algorithm to simultaneous prediction of rock fragmentation and induced ground vibration using unique blast data
,
Applied Sciences
,
12
,
5269
.
Chen
T.
&
Guestrin
C.
(
2016
). '
Xgboost: a scalable tree boosting system
',
Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
, pp.
785
794
.
D'Agostino
V.
&
Ferro
V.
(
2004
)
Scour on alluvial bed downstream of grade-control structures
,
Journal of Hydraulic Engineering
,
130
,
24
37
.
Daneshfaraz
R.
,
Aminvash
E.
,
Di Francesco
S.
,
Najibi
A.
&
Abraham
J.
(
2021a
)
Three-dimensional study of the effect of block roughness geometry on inclined drop
,
Numerical Methods in Civil Engineering
,
6
,
1
9
.
Daneshfaraz
R.
,
Aminvash
E.
,
Ghaderi
A.
,
Abraham
J.
&
Bagherzadeh
M.
(
2021b
)
SVM performance for predicting the effect of horizontal screen diameters on the hydraulic parameters of a vertical drop
,
Applied Sciences
,
11
,
4238
.
Dorogush
A. V.
,
Ershov
V.
&
Gulin
A.
(
2018
)
CatBoost: gradient boosting with categorical features support. arXiv preprint arXiv:1810.11363
.
Edjabou
M. E.
,
Martín-Fernández
J. A.
,
Scheutz
C.
&
Astrup
T. F.
(
2017
)
Statistical analysis of solid waste composition data: arithmetic mean, standard deviation and correlation coefficients
,
Waste Management
,
69
,
13
23
.
Freund
Y.
&
Schapire
R. E.
(
1997
)
A decision-theoretic generalization of on-line learning and an application to boosting
,
Journal of Computer and System Sciences
,
55
,
119
139
.
Gaudio
R.
,
Marion
A.
&
Bovolin
V.
(
2000
)
Morphological effects of bed sills in degrading rivers
,
Journal of Hydraulic Research
,
38
,
89
96
.
Guan
D.
(
2015
)
Scour at submerged weirs. ResearchSpace@ Auckland
.
Hancock
J. T.
&
Khoshgoftaar
T. M.
(
2020
)
Catboost for big data: an interdisciplinary review
,
Journal of Big Data
,
7
,
1
45
.
Hong
J. H.
,
Goyal
M. K.
,
Chiew
Y. M.
&
Chua
L. H.
(
2012
)
Predicting time-dependent pier scour depth with support vector regression
,
Journal of Hydrology
,
468
,
241
248
.
Kalateh
F.
,
Aminvash
E.
&
Abraham
J.
(
2024a
)
On the effect of flow regime and slope of the channel bed on the hydraulic performance of the sharp-crested rectangular side weir: a numerical simulation
,
European Journal of Environmental and Civil Engineering
, 28,
2327
2344
.
Kalateh
F.
,
Aminvash
E.
&
Daneshfaraz
R.
(
2024b
)
On the hydraulic performance of the inclined drops: the effect of downstream macro-roughness elements
,
AQUA – Water Infrastructure, Ecosystems and Society
,
73
,
553
568
.
Khan
S.
,
Waseem
M.
,
Khalid
S.
,
Kontoni
D.-P. N.
,
Ahmad
M.
&
Keawsawasvong
S.
(
2022
)
Fuzzy clustering analysis of HVSR data for seismic microzonation at Lahore city
,
Shock and Vibration
,
2022
,
3109609
.
Lenzi
M. A.
,
Marion
A.
,
Comiti
F.
&
Gaudio
R.
(
2002
)
Local scouring in low and high gradient streams at bed sills
,
Journal of Hydraulic Research
,
40
,
731
739
.
Li
R.
,
Li
D.
,
Yang
J.
,
Xiang
F.
,
Ren
H.
,
Jiang
S.
&
Zhang
L.
(
2021
)
Joint extraction of entities and relations via an entity correlated attention neural model
,
Information Sciences
,
581
,
179
193
.
Li
Q.
,
Diao
H.
,
Ma
Q.
,
Wang
K.
,
Wang
X.
&
Wang
M.
(
2024
)
Geospatial analysis of scour development in offshore wind farms
,
Marine Georesources & Geotechnology
,
1
20
. doi: 10.1080/1064119X.2024.2369945.
Liu
Z.
,
Qiu
H.
,
Zhu
Y.
,
Huangfu
W.
,
Ye
B.
,
Wei
Y.
,
Tang
B.
&
Kamp
U.
(
2024
)
Increasing irrigation-triggered landslide activity caused by intensive farming in deserts on three continents
,
International Journal of Applied Earth Observation and Geoinformation
,
134
,
104242
.
Luo
J.
,
Wang
G.
,
Li
G.
&
Pesce
G.
(
2022
)
Transport infrastructure connectivity and conflict resolution: a machine learning analysis
,
Neural Computing and Applications
,
34
,
6585
6601
.
Meng
S.
,
Zhang
C.
,
Shi
Q.
,
Chen
Z.
,
Hu
W.
&
Lu
F.
(
2023
)
A robust infrared small target detection method jointing multiple information and noise prediction: algorithm and benchmark
,
IEEE Transactions on Geoscience and Remote Sensing
, 61, 1–17.
Muzzammil
M.
,
Alama
J.
&
Danish
M.
(
2015
)
Scour prediction at bridge piers in cohesive bed using gene expression programming
,
Aquatic Procedia
,
4
,
789
796
.
Najafzadeh
M.
,
Etemad-Shahidi
A.
&
Lim
S. Y.
(
2016a
)
Scour prediction in long contractions using ANFIS and SVM
,
Ocean Engineering
,
111
,
128
135
.
Najafzadeh
M.
,
Rezaie Balf
M.
&
Rashedi
E.
(
2016b
)
Prediction of maximum scour depth around piers with debris accumulation using EPR, MT, and GEP models
,
ournal of Hydroinformatics
,
18
,
867
884
.
Parsaie
A.
,
Haghiabi
A. H.
&
Moradinejad
A.
(
2019
)
Prediction of scour depth below river pipeline using support vector machine
,
KSCE Journal of Civil Engineering
,
23
,
2503
2513
.
Prokhorenkova
L.
,
Gusev
G.
,
Vorobev
A.
,
Dorogush
A. V.
&
Gulin
A.
(
2018
)
Catboost: unbiased boosting with categorical features
,
Advances in Neural Information Processing Systems
,
31
, 1–11.
Rashki Ghaleh Nou
M.
,
Azhdary Moghaddam
M.
,
Shafai Bajestan
M.
&
& Azamathulla
H. M.
(
2019
)
Estimation of scour depth around submerged weirs using self-adaptive extreme learning machine
,
Journal of Hydroinformatics
,
21
,
1082
1101
.
Rong
Y.
,
Xu
Z.
,
Liu
J.
,
Liu
H.
,
Ding
J.
,
Liu
X.
,
Luo
W.
,
Zhang
C.
&
Gao
J.
(
2022
)
Du-bus: a realtime bus waiting time estimation system based on multi-source data
,
IEEE Transactions on Intelligent Transportation Systems
,
23
,
24524
24539
.
Schapire
R. E.
(
2013
)
Explaining adaboost
. In:
Empirical Inference: Festschrift in Honor of Vladimir N. Vapnik
,
Berlin, Heidelberg
:
Springer
, pp.
37
52
.
Sharafati
A.
,
Tafarojnoruz
A.
,
Shourian
M.
&
Yaseen
Z. M.
(
2020
)
Simulation of the depth scouring downstream sluice gate: the validation of newly developed data-intelligent models
,
Journal of Hydro-Environment Research
,
29
,
20
30
.
Sharma
C.
&
Ojha
C.
(
2020
). '
Statistical parameters of hydrometeorological variables: standard deviation, SNR, skewness and kurtosis
’,
Advances in Water Resources Engineering and Management: Select Proceedings of TRACE 2018
.
Springer
, pp.
59
70
.
Shi
M.
,
Hu
W.
,
Li
M.
,
Zhang
J.
,
Song
X.
&
Sun
W.
(
2023
)
Ensemble regression based on polynomial regression-based decision tree and its application in the in-situ data of tunnel boring machine
,
Mechanical Systems and Signal Processing
,
188
,
110022
.
Sun
G.
,
Liao
D.
,
Zhao
D.
,
Xu
Z.
&
Yu
H.
(
2015
)
Live migration for multiple correlated virtual machines in cloud-based data centers
,
IEEE Transactions on Services Computing
,
11
,
279
291
.
Sun
G.
,
Song
L.
,
Yu
H.
,
Chang
V.
,
Du
X.
&
Guizani
M.
(
2018a
)
V2V routing in a VANET based on the autoregressive integrated moving average model
,
IEEE Transactions on Vehicular Technology
,
68
,
908
922
.
Sun
G.
,
Zhu
G.
,
Liao
D.
,
Yu
H.
,
Du
X.
&
Guizani
M.
(
2018b
)
Cost-efficient service function chain orchestration for low-latency applications in NFV networks
,
IEEE Systems Journal
,
13
,
3877
3888
.
Tao
H.
,
Habib
M.
,
Aljarah
I.
,
Faris
H.
,
Afan
H. A.
&
Yaseen
Z. M.
(
2021
)
An intelligent evolutionary extreme gradient boosting algorithm development for modeling scour depths under submerged weir
,
Information Sciences
,
570
,
172
184
.
Vanoni
V. A.
(
2006
)
Sedimentation Engineering
.
American Society of Civil Engineers
, Reston, United States.
Wang
E.
,
Yang
Y.
,
Wu
J.
,
Liu
W.
&
Wang
X.
(
2017
)
An efficient prediction-based user recruitment for mobile crowdsensing
,
IEEE Transactions on Mobile Computing
,
17
,
16
28
.
Wang
J.
,
Bai
L.
,
Fang
Z.
,
Han
R.
,
Wang
J.
&
Choi
J.
(
2024
)
Age of information based URLLC transmission for UAVs on pylon turn
,
IEEE Transactions on Vehicular Technology
, 73, 8797–8809.
Wei
W.
,
Xu
W.
,
Deng
J.
&
Guo
Y.
(
2022
)
Self-aeration development and fully cross-sectional air diffusion in high-speed open channel flows
,
Journal of Hydraulic Research
,
60
,
445
459
.
Wei
W.
,
Gong
J.
,
Deng
J.
&
Xu
W.
(
2023
)
Effects of air vent size and location design on air supply efficiency in flood discharge tunnel operations
,
Journal of Hydraulic Engineering
,
149
,
04023050
.
Zhang
X.
,
Hou
D.
,
Xiong
Z.
,
Liu
Y.
,
Wang
S.
&
Li
Y.
(
2024
)
EALLR: energy-aware low-latency routing data driven model in mobile edge computing
,
IEEE Transactions on Consumer Electronics
. doi: 10.1109/TCE.2024.3507158.
Zhou
G.
,
Liu
W.
,
Zhu
Q.
,
Lu
Y.
&
Liu
Y.
(
2022
)
ECA-mobilenetv3 (large)+ SegNet model for binary sugarcane classification of remotely sensed images
,
IEEE Transactions on Geoscience and Remote Sensing
,
60
,
1
15
.
Zi
J.
,
Liu
T.
,
Zhang
W.
,
Pan
X.
,
Ji
H.
&
Zhu
H.
(
2024
)
Quantitatively characterizing sandy soil structure altered by MICP using multi-level thresholding segmentation algorithm
,
Journal of Rock Mechanics and Geotechnical Engineering
, 16, 4285–4299.
Zounemat-Kermani
M.
,
Batelaan
O.
,
Fadaee
M.
&
Hinkelmann
R.
(
2021
)
Ensemble machine learning paradigms in hydrology: a review
,
Journal of Hydrology
,
598
,
126266
.
This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (CC BY 4.0), which permits copying, adaptation and redistribution, provided the original work is properly cited (http://creativecommons.org/licenses/by/4.0/).

Supplementary data