ABSTRACT
Due to the numerous uncertain factors affecting contraction scour depth, although many traditional empirical formulas have been proposed in past research, their prediction accuracy is generally low. In recent years, with advancements in machine learning (ML) technology, these techniques have been able to accurately capture the nonlinear characteristics of scour-depth data. However, in pursuit of higher prediction accuracy, researchers have explored a wide range of diverse ML models that require various combinations of input parameters. These input parameter combinations often lack reliability, and the models themselves have poor interpretability, increasing the ‘black-box effect.’ Therefore, this study uses a principal component analysis (PCA)-enhanced support vector regression (SVR) model to construct a scour depth prediction model, combined with the interpretability method of SHapley Additive exPlanations (SHAP). The results show that the SVR model's predictions are highly consistent with physical experimental laws, and the model primarily identifies features that are strongly linearly correlated with the dependent variable (scour depth and SHAP values). The application of PCA enhances the correlation, and when using the CC-PCA-4 input parameter combination, the SVR model achieves high accuracy (R2 = 0.971, mean absolute percentage error = 7.54%). Moreover, its comprehensive evaluation in terms of stability, accuracy, and conservativeness surpassed that of other ML models and empirical formulas.
HIGHLIGHTS
By using traditional correlation algorithms, the principal component analysis (PCA)-processed data are consistently ranked by the absolute value.
SHAP analysis indicates that the support vector regression model's predictions are consistent with the physical laws of contraction scour.
PCA enhances the linear relationship between independent and dependent variables, thereby improving the overall performance of the model.
INTRODUCTION
In river flow, both artificial and natural structures, such as bridge abutments and landslides, lead to the reduced fluid cross-section area. The reduced width significantly accelerates the flow and therefore leads to increased bed shear stress in the narrowed riverbed, creating a high risk of scour or erosion. Therefore, the accurate calculation of contraction scour depth is crucial for risk assessment and the safe operation of these structures.
Straub (1934) conducted the earliest documented study on contraction scour, deriving a correlation for predicting scour depth under dynamic conditions based on the continuity equations for water flow and sediment. Subsequent expansions and modifications of this correlation were made by researchers (Ashida 1963; Laursen 1963; Komura 1966; Gill 1981; Webby 1984; Lim 1993; Lim & Cheng 1998; Raikar 2004; Briaud et al. 2005; Oliveto & Marino 2019; Nowroozpour & Ettema 2021). Researchers have proposed various criteria for classifying contraction channels. Komura (1966) defined the long contraction channel when the ratio of the contraction section length to the uncontracted channel width (L/b1) exceeds 1, whereas Raikar (2004) and Webby (1984) specified a threshold ratio of 2, above which the channel is regarded as a long contraction channel. Further research indicates that an increase in the length of contraction section reduces and stabilizes shear stress on riverbeds. Additionally, the L/b1 ratio also affects scour morphology, influencing the depth of scour (Oliveto & Marino 2019).
In hydraulic engineering, scouring is generally classified into two types: clear water scour and live-bed scour. Extensive experimental studies using flume tests under controlled laboratory conditions have established a robust set of empirical formulas. Refined by Dey & Raikar (2005), these formulas offer highly accurate predictions for maximum equilibrium scour depth. Despite these advancements, previous studies have often oversimplified the complexity of the problem in their empirical models, limiting their applicability in intricate scenarios (Le et al. 2024). Recently, Lagasse et al. (2021) conducted extensive research and evaluation on the applicability of contraction and pier scour formulas and provided a relatively rare field case study (i.e., the U.S. Highway 287 crossing Spring Creek in Fort Collins, Colorado), marking a critical step toward the practical application of these formulas.
As research progresses, Najafzadeh et al. (2018) expanded the range of predictive methods by using gene expression programming (GEP), M5 Tree (MT), and evolutionary polynomial regression (EPR) to predict maximum equilibrium scour depth. Moreover, enhancing prediction accuracy while reducing the number of input variables has become central to their studies. Najafzadeh et al. (2016) conducted a variable sensitivity analysis using the ANFIS by sequentially removing variables and assessing their impact on model performance, assuming that significant decreases in prediction accuracy would indicate a high importance of the removed variables. More recently, absolute values of correlation coefficients have frequently adopted to assess the importance of variables, combining ML and ensemble techniques to generate numerous hybrid models and select the optimized model that best matches different variable input combinations. For instance, Sharafati et al. (2021) combined five metaheuristic algorithms with the ANFIS model to develop a reliable and robust prediction model that broadly reflects the relationship between input parameters and the target scour depth. The Dagging-Iterative Classifier Optimizer hybrid model by Khosravi et al. (2021) was developed by integrating the Dagging (DA) and Random Subspace (RS) algorithms with five distinct ML models. Recent advances have highlighted the significant role of ML in sediment scour prediction, particularly in estimating scour depth around bridge piers. Eini et al. (2023) provided a comprehensive summary of various ML models and proposed an XGBoost hybrid model enhanced by optimization algorithms, which markedly improved the accuracy of scour depth predictions around circular bridge piers. Furthermore, SHapley Additive exPlanation (SHAP) analysis was employed to interpret the variable importance in the model. In addition, Eini et al. (2024) advanced this approach by combining Bayesian optimization with SVM and XGBoost, achieving enhanced prediction accuracy for scour depth around different pier shapes and demonstrating the substantial potential of ML in scour prediction.
However, the inputs of the aforementioned models are based on traditional correlation measurement methods such as Pearson, Spearman, and Kendall. Constrained by the assumptions (data distribution) of these algorithms, which cause discrepancies in the ranking of variable importance and numerous models with limited interpretability. In recent years, to mitigate the ‘black-box effect’ of prediction models, researchers have adopted interpretable feature selection algorithms such as SHAP or Sobel Operator (SOBEL) analysis for ordering and combining input variables (Guo et al. 2024), given that SHAP results do not rely on any particular model. After identifying variable importance, developing models that can capture these importances becomes a new challenge. Studies on hybrid dimensionality reduction algorithms have confirmed that feature reduction significantly improves the generalization capability and computational efficiency of ML models (Laghrissi et al. 2021; Chang et al. 2022; Pandey et al. 2023). Jia et al. (2022) elaborated on common feature selection and extraction methods and their applications, including PCA, Kernel PCA (KPCA), linear discriminant analysis (LDA), non-negative matrix factorization (NMF), and t-distributed stochastic neighbor embedding (tSNE). They emphasized that during the identification and learning process, adopting appropriate, reliable, and practical feature reduction techniques is crucial for identifying data characteristics and improving accuracy (Cao et al. 2003; Anowar et al. 2021). PCA, as a linear dimensionality reduction technique, transforms high-dimensional data into a lower-dimensional space, effectively mitigating the ‘curse of dimensionality’ by reducing the number of features. Given that the experimental data are sample-based, they are expected to follow the central limit theorem and approximate a Gaussian distribution, aligning well with PCA's assumptions regarding data distribution. This makes PCA especially suitable for handling the high-dimensional, sparse data in this study, as it provides effective denoising, redundancy removal, and computational efficiency. While alternative methods like KPCA and tSNE can also perform dimensionality reduction, they have limitations in data distribution assumptions and feature interpretability, making them less ideal for this study (Jia et al. 2022). Additionally, PCA offers a scalable approach for handling additional data features that may emerge in future observational datasets. To further capture the nonlinear relationships in the data post-dimension reduction, we selected support vector regression (SVR) with a Gaussian kernel as the ML model. SVR leverages kernel functions to map data into a higher-dimensional feature space, enabling the accurate capture of nonlinear patterns with greater precision and robustness. With fewer hyperparameters, it is easier to optimize and, due to its inherent regularization, can avoid overfitting in limited samples, supporting improved generalizability when expanding the dataset in the future (Eini et al. 2024).
Therefore, we developed an interpretable scour depth prediction model based on PCA dimensionality reduction and SVR. The traditional variable importance rankings were unified using PCA algorithm, and the SVR model was optimized through Bayesian grid search and k-fold cross-validation to learn the features of data. Meanwhile, SHAP was used to reveal the patterns learned by the model from the dataset. Additionally, to comprehensively evaluate the predictive performance of the optimized model, it was compared with other literature models in terms of metrics and uncertainty analysis, and its accuracy and conservatism were compared with empirical formulas.
DIMENSIONAL ANALYSIS AND DATASET PREPARATION
Table 1 summarizes the distribution of available datasets from contraction channel clear water scour experiments (Laursen 1963; Komura 1966; Gill 1981; Webby 1984; Lim 1993; Dey & Raikar 2005) in literature. The summary includes additional details not found in earlier literature, such as the contraction channel length (L) for each of the 182 experimental samples (Lagasse et al. 2021).
Parameters . | Stage . | |
---|---|---|
Training . | Testing . | |
L0 | 1.6667-10 | 1.6667-10 |
d0 | 8.75 × 10−4–0.02375 | 8.75 × 10−4–0.02375 |
Fr0 | 1.1434–3.3497 | 1.6081–3.3497 |
h0 | 0.0360–0.2288 | 0.0509–0.2277 |
b0 | 0.25–0.7 | 0.25–0.7 |
σg | 1.065–3.6 | 1.08–3.6 |
S | 2.59–2.65 | 2.59–2.928 |
U1/UC | 0.392–1 | 0.517–0.974 |
ds0 | 0.0132–0.2483 | 0.0267–0.2567 |
Parameters . | Stage . | |
---|---|---|
Training . | Testing . | |
L0 | 1.6667-10 | 1.6667-10 |
d0 | 8.75 × 10−4–0.02375 | 8.75 × 10−4–0.02375 |
Fr0 | 1.1434–3.3497 | 1.6081–3.3497 |
h0 | 0.0360–0.2288 | 0.0509–0.2277 |
b0 | 0.25–0.7 | 0.25–0.7 |
σg | 1.065–3.6 | 1.08–3.6 |
S | 2.59–2.65 | 2.59–2.928 |
U1/UC | 0.392–1 | 0.517–0.974 |
ds0 | 0.0132–0.2483 | 0.0267–0.2567 |
Note: Lim's data are based on the portions listed by Dey & Raikar (2005). All data have been reviewed and confirmed to have been measured under dynamic equilibrium conditions and to be free of scale effects (Laursen 1963; Komura 1966; Gill 1981; Webby 1984; Lim 1993; Dey & Raikar 2005).
METHODOLOGY
Modeling method steps
Dimensionality reduction algorithms
Principal component analysis
Other dimensionality reduction algorithms
In addition to PCA, several other dimensionality reduction techniques are utilized to handle complex data structures. These include KPCA, tSNE, LDA, and NMF. (1) KPCA extends PCA to nonlinear data structures by using a Kernel trick to project data into a higher-dimensional space, enabling the capture of complex patterns (Schölkopf et al. 1997); (2) tSNE excels in visualizing high-dimensional data by preserving local relationships through probabilistic distributions, making it suitable for exploratory data analysis (Gisbrecht et al. 2015); (3) LDA differentiates itself by focusing on maximizing class separability, which is highly beneficial in supervised learning scenarios where class labels are known (Park & Park 2008); (4) NMF is particularly used in contexts like image processing and text mining, where it decomposes non-negative data into simpler, meaningful components, facilitating topic identification and feature extraction (Lee & Seung 1999).
The parameter settings for the five dimensionality reduction algorithms mentioned above are detailed in Table 2.
Preprocessing . | Configure . |
---|---|
PCA | Normalization |
Covariance matrix | |
KPCA | Gaussian kernel = 1 |
tSNE | Perplexity = 30 |
Algorithm = exact | |
Exaggeration = 4 | |
Learning rate = 200 | |
LDA | Discrimination type = linear |
NMF | Algorithm = alternating least squares |
Convergence threshold = 10−4 |
Preprocessing . | Configure . |
---|---|
PCA | Normalization |
Covariance matrix | |
KPCA | Gaussian kernel = 1 |
tSNE | Perplexity = 30 |
Algorithm = exact | |
Exaggeration = 4 | |
Learning rate = 200 | |
LDA | Discrimination type = linear |
NMF | Algorithm = alternating least squares |
Convergence threshold = 10−4 |
Support vector regression
SHAP analysis
SHAP is an advanced model interpretation method used to address the ‘black-box effect’ in ML models. It provides an explanation for model decisions by calculating the average marginal contribution of each feature to the model prediction, known as the Shapley value. The sign of the Shapley value (positive or negative) indicates the direction in which the feature affects the prediction result, while the magnitude of the absolute value reflects the importance of the feature: the larger the absolute value, the more significant the influence of the feature, and thus the more important the feature is considered. The application of SHAP has expanded to various technical and engineering fields, including the scour domain (Kim et al. 2024), demonstrating its broad applicability and practical value.
Model evaluation
To evaluate the model performances, eight evaluation metrics are used: coefficient of determination (R2), root mean square error (RMSE), mean absolute error (MAE), mean bias error (MBE), mean absolute percentage error (MAPE), Pearson's correlation coefficient (CC) (Kim et al. 2024; Kumar et al. 2024).
RESULTS AND DISCUSSION
Correlation ranking and interpretation
Table 3 summarizes the relative sensitivity of the scour depth for different input variables derived based on different criteria in the literature. Raikar et al. (2016) used PCA, selecting the factors with the largest proportion of principal components (PCs) as representative variables to explain the results of the model (e.g., PC1 = F1ec). However, since each PC is composed of a linear combination of all original dimensionless quantities, it is difficult to distinguish the dominant variables. Najafzadeh et al. (2016) evaluated the sensitivity of the results to the input variables by sequentially excluding each variable from the ANFIS model ensemble. A highly important variable is recognized as its removal leads to a significant drop in model prediction accuracy. Although this method provides a preliminary assessment, the mechanisms by which variables influence the outcomes remain unexplained, and the interpretability is highly dependent on the selected ML model. Though not explicitly stated, it is inferred that Sharafati et al. (2021) used the Pearson correlation method (Table 3 – current investigation before PCA) and observed that the shrinkage ratio b0 is the most important factor, exhibiting a negative correlation with the results. This observation is consistent with previous studies, i.e., the larger the b0 = b2/b1, the less concentrated the flow and the smaller the scouring capacity. Khosravi et al. (2021) also found that b0 is negatively correlated with the scour depth, but their results showed that σg is positively correlated with the scour depth, contrary to the known ‘shielding effect’ theory, which suggests that an increase in the sediment non-uniformity coefficient should reduce scour depth. Previous studies have consistently found that h0 is always positively correlated with scour depth, in accordance with experiments suggesting that larger h0 values are more likely to generate large-scale vortex structures, thereby increasing scour depth. However, the importance of h0 varies across different studies, and no uniform significance has been assigned. This variation could be attributed to changes in h0 during actual scour processes, as noted by Lagasse et al. (2021).
. | Previous research . | |||||||
---|---|---|---|---|---|---|---|---|
Khosravi et al. (2021) | Fr0 | b0 | h0 | σg | U1/UC | d0 | – | – |
CC | −0.624 | −0.578 | 0.563 | 0.331 | 0.33 | 0.292 | – | – |
Sharafati et al. (2021) | b0 | h0 | σg | U1/UC | d0 | Fr0 | S | – |
CC | −0.578 | 0.347 | −0.271 | 0.263 | 0.259 | 0.124 | 0 | – |
Najafzadeh et al. (2016) | h0 | d0 | b0 | σg | Fr0 | U1/UC | – | – |
Sensitivity by ANFIS | 0.6 | 0.84 | 0.84 | 0.85 | 0.85 | 0.91 | – | – |
Raikar et al. (2016) | F1ec(PC1) | d0 (PC2) | b0 (PC3) | h0 (PC4) | Fr0 (PC5) | – | – | – |
Variance by PCA | 38.7% | 28.1% | 22.7% | 8.3% | 2.2% | – | – | – |
. | Current investigation before PCA . | |||||||
Parameter | b0 | h0 | σg | U1/UC | d0 | S | Fr0 | L0 |
CC of Pearson | −0.578 | 0.343 | −0.329 | 0.279 | 0.26 | −0.229 | 0.125 | 0.117 |
Parameter | b0 | σg | d0 | L0 | h0 | S | U1/UC | Fr0 |
CC of Spearman | −0.615 | −0.308 | 0.299 | −0.293 | 0.288 | −0.25 | 0.158 | 0.062 |
Parameter | b0 | L0 | σg | d0 | h0 | S | U1/UC | Fr0 |
CC of Kendall | −0.484 | −0.239 | −0.226 | 0.214 | 0.204 | −0.203 | 0.103 | 0.039 |
SHAP | b0 | h0 | σg | Fr0 | d0 | U1/UC | S | L0 |
SHAP values | 0.029 | 0.016 | 0.012 | 0.009 | 0.008 | 0.004 | 0.003 | 0.002 |
. | Current investigation after PCA . | |||||||
Parameter | PC3 | PC5 | PC4 | PC8 | PC7 | PC2 | PC6 | PC1 |
CC of Pearson | 0.649 | 0.628 | 0.128 | 0.099 | 0.078 | 0.072 | 0.049 | 0.047 |
CC of Spearman | 0.705 | 0.631 | 0.183 | 0.174 | 0.098 | 0.079 | 0.063 | 0.029 |
CC of Kendall | 0.523 | 0.458 | 0.161 | 0.136 | 0.063 | 0.051 | 0.046 | 0.031 |
Parameter | PC5 | PC3 | PC1 | PC4 | PC6 | PC2 | PC8 | PC7 |
SHAP values | 0.031 | 0.029 | 0.007 | 0.005 | 0.004 | 0.004 | 0.003 | 0.001 |
Parameter | PC1 | PC2 | PC3 | PC4 | PC5 | PC6 | PC7 | PC8 |
Variance by PCA | 33.1% | 27.5% | 14.5% | 11.4% | 8.3% | 4.9% | 0.2% | 0.1% |
. | Previous research . | |||||||
---|---|---|---|---|---|---|---|---|
Khosravi et al. (2021) | Fr0 | b0 | h0 | σg | U1/UC | d0 | – | – |
CC | −0.624 | −0.578 | 0.563 | 0.331 | 0.33 | 0.292 | – | – |
Sharafati et al. (2021) | b0 | h0 | σg | U1/UC | d0 | Fr0 | S | – |
CC | −0.578 | 0.347 | −0.271 | 0.263 | 0.259 | 0.124 | 0 | – |
Najafzadeh et al. (2016) | h0 | d0 | b0 | σg | Fr0 | U1/UC | – | – |
Sensitivity by ANFIS | 0.6 | 0.84 | 0.84 | 0.85 | 0.85 | 0.91 | – | – |
Raikar et al. (2016) | F1ec(PC1) | d0 (PC2) | b0 (PC3) | h0 (PC4) | Fr0 (PC5) | – | – | – |
Variance by PCA | 38.7% | 28.1% | 22.7% | 8.3% | 2.2% | – | – | – |
. | Current investigation before PCA . | |||||||
Parameter | b0 | h0 | σg | U1/UC | d0 | S | Fr0 | L0 |
CC of Pearson | −0.578 | 0.343 | −0.329 | 0.279 | 0.26 | −0.229 | 0.125 | 0.117 |
Parameter | b0 | σg | d0 | L0 | h0 | S | U1/UC | Fr0 |
CC of Spearman | −0.615 | −0.308 | 0.299 | −0.293 | 0.288 | −0.25 | 0.158 | 0.062 |
Parameter | b0 | L0 | σg | d0 | h0 | S | U1/UC | Fr0 |
CC of Kendall | −0.484 | −0.239 | −0.226 | 0.214 | 0.204 | −0.203 | 0.103 | 0.039 |
SHAP | b0 | h0 | σg | Fr0 | d0 | U1/UC | S | L0 |
SHAP values | 0.029 | 0.016 | 0.012 | 0.009 | 0.008 | 0.004 | 0.003 | 0.002 |
. | Current investigation after PCA . | |||||||
Parameter | PC3 | PC5 | PC4 | PC8 | PC7 | PC2 | PC6 | PC1 |
CC of Pearson | 0.649 | 0.628 | 0.128 | 0.099 | 0.078 | 0.072 | 0.049 | 0.047 |
CC of Spearman | 0.705 | 0.631 | 0.183 | 0.174 | 0.098 | 0.079 | 0.063 | 0.029 |
CC of Kendall | 0.523 | 0.458 | 0.161 | 0.136 | 0.063 | 0.051 | 0.046 | 0.031 |
Parameter | PC5 | PC3 | PC1 | PC4 | PC6 | PC2 | PC8 | PC7 |
SHAP values | 0.031 | 0.029 | 0.007 | 0.005 | 0.004 | 0.004 | 0.003 | 0.001 |
Parameter | PC1 | PC2 | PC3 | PC4 | PC5 | PC6 | PC7 | PC8 |
Variance by PCA | 33.1% | 27.5% | 14.5% | 11.4% | 8.3% | 4.9% | 0.2% | 0.1% |
Given the discrepancy in ranking order and the lack of interpretability, which might increase the difficulty of matching with appropriate ML models, it is necessary to carefully review the consistency of different commonly used variable correlation algorithms in identifying variable importance before inputting them into ML models. In the ‘current investigation before PCA’ section of Table 3, the relative significances indicated by the correlation coefficients of different input variables derived from Pearson, Spearman, and Kendall algorithms are listed. Except for b0, which is consistently observed to posses a negative correlation coefficient with the highest magnitude. The relative significances of the remaining variables vary among different models. This phenomenon may stem from the assumptions of traditional Pearson, Spearman, and Kendall algorithms, which, respectively, require linear relationships, monotonic relationships, and association consistency. These assumptions are difficult to satisfy for datasets with small sample sizes.
As shown in the ‘Current investigation after PCA’ section of Table 3, after PCA processing, the importance rankings of traditional correlation algorithms have been unified, with the importance of PCs decreasing in the order of PC3, PC5, PC4, PC8, PC7, PC2, PC6, and PC1. Although the SHAP rankings of PCs show slight inconsistencies, three of the top four rankings are consistent with traditional algorithms, which might be due to the use of the RBF kernel in SVR.
For h0, its SHAP value nearly linearly increases from −0.04 to 0.02 as the feature values increase, which aligns with observed differences in wall shear stress contributions between shallow water effects at lower depths and the formation of large-scale vortices, such as horseshoe or necklace vortices, at greater depths. The development of these vortex structures is a significant factor in increasing scour depth (Eini et al. 2024), and current data indicate that this trend shows no signs of diminishing. The increase in particle gradation σg stabilizes its impact on scour depth, reducing scour depth, which aligns with actual physical processes. The impact of Fr0 on scour depth is similar to that of h0, with a linear increase, following the mechanism described by Raikar et al. (2016) using excess Froude number (F1ec), where Froude number comprises the difference between the contraction flow velocity and the critical sediment entrainment velocity. When this value exceeds 0, it significantly increases scour depth; however, current data only involve clear water scour, and more data are required for live-bed scour. The slight decrease of d0 near zero is consistent with Dey & Raikar's (2005) experimental findings, which are due to variations in the bed shear stress requirements caused by the transitional properties of the Shields curve. For gravel, in order to maintain the flow velocity U1 close to UC, U1 was significantly increased (thus increasing U1/UC), greatly enhancing bed shear stress in the contraction zone and thereby increasing scour potential, leading ds0 to increase at a constant rate. Although the sample size for density ratio S is limited, its negative correlation with scour depth aligns with experimental laws. Before PCA processing, SHAP visualization analysis reveals significant linear relationships between the top-ranked variables such as b0 (CC = −0.933), h0 (CC = 0.952), and σg (CC = −0.952) and the SHAP values of the dependent variable (Figure 4(a)), with the trend of influence consistent with the physical interpretation (Section 4.1). In contrast, lower-ranked variables S and L0 did not show clear linear relationships. (Previous research indicates that high S values reduce sediment scour, but most natural sediments cluster around an S value of 2.65, suggesting a need for more diverse data.) Additionally, longer L values may alter flow control from downstream to upstream, a hypothesis that requires further validation with larger datasets (Lagasse et al. 2021). The above observations indicate a high correlation between the top-ranked variables identified by traditional correlation algorithms and their SHAP values in the ‘current investigation after PCA’ section of Table 3.
To further prove this point, when the data are processed by PCA (as shown in Figure 4(d)), SHAP visualization analysis shows that as variables such as PC3 (CC = 0.977) and PC5 (CC = 0.994) increased, the corresponding SHAP values exhibit significant linear growth. These variables consistently ranked high in traditional algorithms, indicating a strong linear contribution relationship between PC3, PC5, and scour depth. This indirectly confirms that the PCA method's effect of linearizing data features make it easier for SVR to identify these features. To directly verify this point, Section 4.2 will input variables based on correlation rankings to explore their specific impact on prediction accuracy.
Performance of the model
The method of cumulatively inputting variables ranked by the conventional PCA principal component variance contribution rate did not yield ideal results (as shown in Figure 5(c)), reaching the maximum R2 = 0.936 when five PCs were input. However, adding more PCs has a negligible effect on the accuracy. This is mainly due to that PC3, PC4, and PC5 have been included in the first five PCs.
If PCA is not performed and the dimensionless variables ranked by SHAP analysis are directly input in order of importance, an R2 = 0.955 can be achieved with only b0, h0, and σg taken as inputs. However, as more variables are added, the accuracy slightly decreases, confirming the physical experimental laws that adding more physical variables does not necessarily lead to better results (e.g., ignoring the Reynolds number in dimensionless analysis because it is not considered in turbulent flows). Most studies also overlook d0 and U1/UC when fitting data using empirical formulas, as outlined in Section 4.3. Nevertheless, it can at least enhance the stability of SVR predictions. Moreover, the SHAP interpretability analysis method's SHAP value distribution for scour depth is consistent with the conventional theoretical knowledge of scour depth influencing factors, indicating that the SVR model's predictions are reasonable (the Laursen formula in Section 4.3. also uses b0 and h0 variables).
Further uncertainty analysis (see Table 4) revealed that all models tend to underestimate (negative MBE). Additionally, the standard deviation is defined as where is the predicted value and is the observed value, and the 95% confidence interval (95% CI) is calculated to provide precise estimates (Ebtehaj et al. 2017). The results show that the CC-PCA-4 model has the smallest and narrowest uncertainty interval ([−0.01816, 0.01483]), further confirming its superior predictive performance.
Model input . | MBE . | Se . | Bandwidth . | 95%CI . |
---|---|---|---|---|
CC-PCA-4 | −0.00167 | 0.008415814 | 0.032989 | [−0.01816, 0.01483] |
SHAP-PCA-4 | −0.00639 | 0.008822983 | 0.034585 | [−0.02368, 0.01091] |
Cumulative-5 | −0.00458 | 0.009129544 | 0.035787 | [−0.02247, 0.01331] |
SHAP-3 | −0.00112 | 0.008843427 | 0.034666 | [−0.01845, 0.01622] |
Model input . | MBE . | Se . | Bandwidth . | 95%CI . |
---|---|---|---|---|
CC-PCA-4 | −0.00167 | 0.008415814 | 0.032989 | [−0.01816, 0.01483] |
SHAP-PCA-4 | −0.00639 | 0.008822983 | 0.034585 | [−0.02368, 0.01091] |
Cumulative-5 | −0.00458 | 0.009129544 | 0.035787 | [−0.02247, 0.01331] |
SHAP-3 | −0.00112 | 0.008843427 | 0.034666 | [−0.01845, 0.01622] |
To provide a more comprehensive evaluation of the model, we have added a sensitivity analysis. The sensitivity of the input variables to the output was analyzed using the best SVR model. Multiple training datasets were generated by removing each input variable one by one, and results were evaluated based on the R2 metric. Findings in Table 5 indicate that b0 plays a critical role in predicting scour depth compared to other input variables, as it yielded the lowest R2 on the test set. For σg and h0; however, the sensitivity ranking differs slightly from the importance derived through SHAP analysis. This is understandable, as both variables are similarly close in terms of sensitivity and importance. While SHAP values measure each feature's marginal contribution to the model output, eliminating a feature during hyperparameter optimization changes the interactions among the remaining features. These complex interactions may lead to different sensitivity analysis results. For the other variables, the sensitivity ranking is consistent with the SHAP analysis (Figure 4(b)).
Input combinations . | Parameters to be removed . | R2 (training) . | R2 (testing) . |
---|---|---|---|
h0, σg, Fr0, d0, U1/UC, S, L0 | b0 | 0.4656 | 0.2857 |
b0, σg, Fr0, d0, U1/UC, S, L0 | h0 | 0.9721 | 0.8267 |
b0, h0, Fr0, d0, U1/UC, S, L0 | σg | 0.8355 | 0.8191 |
b0, h0, σg, d0, U1/UC, S, L0 | Fr0 | 0.9050 | 0.8480 |
b0, h0, σg, Fr0, U1/UC, S, L0 | d0 | 0.9429 | 0.8847 |
b0, h0, σg, Fr0, d0, S, L0 | U1/UC | 0.9696 | 0.9218 |
b0, h0, σg, Fr0, d0, U1/UC, L0 | S | 0.9803 | 0.9259 |
b0, h0, σg, Fr0, d0, U1/UC, S | L0 | 0.9843 | 0.9524 |
Input combinations . | Parameters to be removed . | R2 (training) . | R2 (testing) . |
---|---|---|---|
h0, σg, Fr0, d0, U1/UC, S, L0 | b0 | 0.4656 | 0.2857 |
b0, σg, Fr0, d0, U1/UC, S, L0 | h0 | 0.9721 | 0.8267 |
b0, h0, Fr0, d0, U1/UC, S, L0 | σg | 0.8355 | 0.8191 |
b0, h0, σg, d0, U1/UC, S, L0 | Fr0 | 0.9050 | 0.8480 |
b0, h0, σg, Fr0, U1/UC, S, L0 | d0 | 0.9429 | 0.8847 |
b0, h0, σg, Fr0, d0, S, L0 | U1/UC | 0.9696 | 0.9218 |
b0, h0, σg, Fr0, d0, U1/UC, L0 | S | 0.9803 | 0.9259 |
b0, h0, σg, Fr0, d0, U1/UC, S | L0 | 0.9843 | 0.9524 |
Reference . | Model . | CC . | R2 . | RMSE . | MAPE (%) . |
---|---|---|---|---|---|
Current | CC-PCA-4-SVR | 0.9863 | 0.971 | 0.0104 | 7.54 |
CC-PCA-6-BP | 0.9368 | 0.8709 | 0.0181 | 18.61 | |
CC-PCA-7-GA-BP | 0.9475 | 0.8957 | 0.0162 | 11.2 | |
CC-PCA-5-PSO-BP | 0.9402 | 0.8808 | 0.0174 | 12.74 | |
CC-PCA-6-ELM | 0.9524 | 0.905 | 0.0155 | 14.57 | |
CC-PCA-6-RBF | 0.9666 | 0.9286 | 0.0134 | 11.11 | |
Khosravi et al. (2021) | DA-ICO-4 | 0.95 | – | 0.01 | – |
Sharafati et al. (2021) | ANFIS-BBO-M2 | 0.9619 | – | 0.0158 | – |
Najafzadeh et al. (2018) | EPR | 0.9030 | 0.903 | 0.0263 | – |
Najafzadeh et al. (2016) | ANFIS | 0.8900 | – | 0.0281 | 27.54 |
Raikar et al. (2016) | GA-3 | – | 0.955 | – | 13.2 |
Laursen (1963) | Equation (5) | 0.82 | 0.6724 | 0.0211 | 27.58 |
Komura (1966) | Equation (6) | 0.7944 | 0.631 | 0.0889 | 141.09 |
Gill (1981) | Equation (7) | 0.8226 | 0.6766 | 0.123 | 189.44 |
Lim (1993) | Equation (8) | 0.8295 | 0.6881 | 0.0937 | 144.94 |
Reference . | Model . | CC . | R2 . | RMSE . | MAPE (%) . |
---|---|---|---|---|---|
Current | CC-PCA-4-SVR | 0.9863 | 0.971 | 0.0104 | 7.54 |
CC-PCA-6-BP | 0.9368 | 0.8709 | 0.0181 | 18.61 | |
CC-PCA-7-GA-BP | 0.9475 | 0.8957 | 0.0162 | 11.2 | |
CC-PCA-5-PSO-BP | 0.9402 | 0.8808 | 0.0174 | 12.74 | |
CC-PCA-6-ELM | 0.9524 | 0.905 | 0.0155 | 14.57 | |
CC-PCA-6-RBF | 0.9666 | 0.9286 | 0.0134 | 11.11 | |
Khosravi et al. (2021) | DA-ICO-4 | 0.95 | – | 0.01 | – |
Sharafati et al. (2021) | ANFIS-BBO-M2 | 0.9619 | – | 0.0158 | – |
Najafzadeh et al. (2018) | EPR | 0.9030 | 0.903 | 0.0263 | – |
Najafzadeh et al. (2016) | ANFIS | 0.8900 | – | 0.0281 | 27.54 |
Raikar et al. (2016) | GA-3 | – | 0.955 | – | 13.2 |
Laursen (1963) | Equation (5) | 0.82 | 0.6724 | 0.0211 | 27.58 |
Komura (1966) | Equation (6) | 0.7944 | 0.631 | 0.0889 | 141.09 |
Gill (1981) | Equation (7) | 0.8226 | 0.6766 | 0.123 | 189.44 |
Lim (1993) | Equation (8) | 0.8295 | 0.6881 | 0.0937 | 144.94 |
The bold values emphasize that this model is optimal.
Comparison with previous studies
Khosravi et al. (2021) evaluated five models – isotonic regression (ISOR), sequential minimal optimization (SMO), iterative classifier optimizer (ICO), locally weighted learning (LWL), and least median of squares regression (LMS) – combined with the DA and RS optimization algorithms, finding that the DA-ICO-4 model performed best. Sharafati et al. (2021) applied five optimization algorithms – ant colony optimization (ACO), biogeographic-based optimization (BBO), GA, invasive weeds optimization (IWO), and teaching–learning-based optimization (TLBO) – to optimize the ANFIS model, with ANFIS-BBO-M2 showing the best performance. Najafzadeh et al. (2016) used a base version of ANFIS, which outperformed a simple hyperparameter-tuned SVM (CC = 0.88); they subsequently applied GEP (CC = 0.89), MT (CC = 0.874), and EPR (CC = 0.903), with EPR yielding relatively better results. Raikar et al. (2016) tested ANN (BP) models with single and double hidden layers, as well as GA. Pursuing a lower MSE, the single-layer network (4-16-1 structure) achieved an R2 of only 0.808, inferior to the GA's R2 of 0.955.
CONCLUSIONS
This research proposes a PCA-enhanced SVR model to predict contraction scour depth. Around 182 laboratory samples were used to construct the model. SHAP analysis was used to explain the prediction mechanism of the model, and its effectiveness was evaluated by comparison with a priori knowledge. The use of PCA allows the ML model to achieve high accuracy without the need for largely extensive searches for input variable matching. Finally, the model's predictive performance was validated by comparing it with existing models and empirical formulas from the aspects of uncertainty, predictive distribution, accuracy, and conservatism. Overall, this study validates the feasibility of using the interpretable ML approach to predict scour depth and provides an explanation and comprehensive evaluation of its mechanism. The detailed conclusions of this study are as follows:
1. SHAP analysis shows that the SVR model tends to identify variables that are strongly linearly correlated with scour depth and SHAP values. The identified key variables, such as b0, h0, and σg, have influence trends consistent with the physical experimental laws.
2. The data after PCA dimensionality reduction enhanced the correlation between independent variables (PCs) and dependent variables (scour depth and SHAP values), making the importance rankings of PCA-processed variables more consistent according to traditional correlation algorithms.
3. The SVR model outperforms other models and empirical formulas in accuracy. With the top four PCA-processed correlated input parameters (CC-PCA-4 scheme), the R2 value is 0.971, and MAPE is 7.54%. Its results show a narrower uncertainty interval ([–0.01816, 0.01483]). Conservatism analysis shows that when the safety factor k = 1.2, the model's prediction error (MAPE) is 22.1% while maintaining a high conservatism of 94.6%.
Despite the effectiveness of the model demonstrated in this study, there remain certain limitations regarding data scale and complexity. Data acquisition costs are exceedingly high, whether from physical sediment model experiments, numerical simulations, or field measurements. This constraint is one reason why current research predominantly relies on laboratory data. Given these conditions, a feasible approach for rapid dataset expansion is to establish an automated experimental and monitoring platform to address data gaps in density ratios and contraction segment lengths. Additionally, we aim to integrate more statistical metrics into the existing data types through this platform, such as time-varying turbulence energy, sediment concentration, and time-varying scour depth, thereby extending the data range to include live-bed scour and extreme flood conditions. We also plan to collect more field measurement cases and, based on this enriched dataset, develop a physics-based online support vector machine that can enable dynamic prediction, assessment, and early warning, providing a reliable tool for infrastructure safety and health monitoring.
ETHICS APPROVAL
The authors are consistent with the ethical requirements.
CONSENT TO PARTICIPATE
The authors all consent to participate in the paper editing.
CONSENT FOR PUBLICATION
The authors all consent to the publication of the paper.
FUNDING
The authors wish to acknowledge the financial support by the National Natural Science Foundation of China (Grant Nos. 52179060, 52209081, and 52479060).
DATA AVAILABILITY STATEMENT
Data cannot be made publicly available; readers should contact the corresponding author for details.
CONFLICT OF INTEREST
The authors declare there is no conflict.