Precipitation nowcasting plays a crucial role in disaster prevention and mitigation. Existing forecasting models often underutilize output data, leading to suboptimal forecasting performance. To tackle this issue, we introduce the I-ConvGRU model, a novel radar echo timing prediction model that synergizes the temporal dynamics optimization of ConvGRU with the spatial feature enhancement capabilities of RainNet. The model forecasts future scenarios by processing 10 sequential time-series images as input while employing skip connections to boost its spatial feature representation further. Evaluation of the radar echo data set from the Hong Kong Hydrological and Meteorological Bureau spanning from 2009 to 2015 demonstrates the I-ConvGRU model's superiority, with reductions of 17(3.8%) and 49(3.2%) in MSE and MAE metrics, respectively, compared with the TrajGRU model; meanwhile, the I-ConvGRU model had 52(5.8%) and 144(3.8%) lower values on the B-MSE and B-MAE metrics, respectively, than the slightly better performing TrajGRU model. Notably, it significantly improves the prediction of severe precipitation events, with the CSI and HSS metrics increasing by 0.0251(9.6%) and 0.0277(6.8%). These results affirm the model's enhanced effectiveness in radar echo forecasting, particularly in predicting heavy rainfall events.

  • The model designed by combining recurrent neural networks and iterative networks improves the warning rate of radar echo extrapolation.

  • The model has strong early warning ability for heavy rainfall.

Severe convective weather is characterized by its limited spatial extent, brief duration, intense unpredictability, and significant destructive potential (Wang et al. 2022; Chowdhury et al. 2023; Xiang et al. 2023). It often triggers catastrophic conditions, including thunderstorms, hail, strong winds, and heavy rainfall (Zhang & Melhauser 2012). Additionally, it represents a significant natural hazard in China (Zhang & Melhauser 2012) and it is also a key challenging aspect in current meteorological forecasting operations (Yan et al. 2022). While conventional numerical meteorological models, enhanced by high-performance computing and data assimilation techniques, have markedly improved weather forecasting accuracy (Bauer et al. 2015), the computational demands for predicting precipitation dynamics at a high spatiotemporal resolution are generally not well suited for real-time nowcasting operations, which require frequent updates (every 5–10 min). Furthermore, due to their inherently nonlinear characteristics, traditional dynamic framework-based prediction methods struggle to forecast severe convective weather accurately.

Forecasting severe convective weather is typically categorized into short-term forecasts and nowcasts. A short-term forecast covers 0–12 h, focusing primarily on the potential for extreme weather events within this timeframe. Conversely, nowcasting, which spans a shorter lead time of up to 2 h (Shukla et al. 2014), emphasizes immediate responses to imminent weather changes. Weather radar serves as a pivotal tool in nowcasting, offering an unparalleled spatial and temporal resolution for tracking precipitation systems and forecasting medium to small-scale severe convective events (Wang et al. 2023). Currently, radar echo extrapolation is the principal technique in nowcasting. This method extrapolates the future location and intensity of radar echoes based on current weather radar observations to predict severe convective weather (Rinehart & Garvey 1978; Chen et al. 2022). Many studies have confirmed that heuristic extrapolation of precipitation dynamics, as observed by weather radar, consistently surpasses the accuracy of numerical forecast models for short lead times (Lin et al. 2005; Yin et al. 2021). With the ongoing expansion and operation of weather radar networks, advancements in radar echo extrapolation algorithms leveraging radar observation data are crucial for minimizing the impact of meteorological disasters (Shi et al. 2018).

Current research in echo extrapolation predominantly utilizes the single-body centroid, cross-correlation, and optical flow methods. The single-body centroid method, rooted in three-dimensional thunderstorm tracking technology (Sun et al. 2014), identifies and analyzes single-body features within the echo to predict the echo's future position. This approach has proven effective for tracking monomers with strong echoes and compact volumes, and various algorithms based on this method have seen widespread application (Dixon & Wiener 1993; Johnson et al. 1998; Dai et al. 2022). The cross-correlation method, or tracking radar echoes by correlation (TREC) (Zou et al. 2019), leverages optimal spatial correlations of echoes to trace precipitation system movements. It accounts for echo size, movement direction, and deformations occurring during transit, achieving notable success in nowcasting operations (Li et al. 1995, 2013). However, each method has its constraints: the single-body centroid method is best suited for convective severe thunderstorm cells and less effective in general convective weather forecasting. The cross-correlation method excels in slowly changing stratiform cloud precipitation systems but falters with rapidly changing precipitation echoes, as rapid weather changes introduce significant errors in motion vector field calculations, diminishing forecast accuracy (Zhu & Dai 2022). The optical flow method, extensively applied in practice, originates from computer vision. A notable variant is the variational optical flow method, ROVER (Woo & Wong 2017), which calculates the optical flow field of consecutive radar images based on a stationary assumption, employing the technique suggested by Brox et al. (2004) and implementing the semi-Lagrangian advection scheme for precipitation forecasting (Tishchenko et al. 2019). Although the optical flow method has an inherent physical change mechanism, the separation between calculation and extrapolation complicates the determination of model parameters, presenting challenges in its application.

In recent years, the advent of large-scale parallel computing and the widespread adoption of graphics processing unit (GPU) devices have significantly enhanced computing capabilities. Following the establishment of a new generation of weather radar networks in China (Min et al. 2019), there has been an exponential increase in the volume of data available for training models. Consequently, deep learning has increasingly been applied to radar echo extrapolation. With its robust feature recognition capabilities, deep learning can discern complex patterns and extract meaningful insights from radar echo data, leading to more accurate extrapolation predictions. Additionally, deep learning models excel in modeling complex nonlinear relationships inherent in radar echo data, enhancing forecasts' precision and reliability (Zhang et al. 2022). The most notable deep learning models recently deployed in this field fall into three main categories: recurrent neural networks (RNNs), convolutional neural networks (CNNs), and generative adversarial networks (GANs).

In the realm of radar echo extrapolation, RNNs have seen significant application, with long short-term memory networks (LSTMs) and gated recurrent units (GRUs) being particularly prominent. Shi et al. (2015) introduced the ConvLSTM model, which integrates a convolution layer and a sampling layer to extract spatial characteristics of sequences and accurately capture the spatiotemporal dynamics of radar echo patterns. However, due to its complexity, Shi et al. (2017) later proposed more streamlined architectures, namely the Trajectory Gated Recurrent Unit (TrajGRU) and Convolution Gated Recurrent Unit (ConvGRU). The TrajGRU model, which accounts for motion trajectories, achieves superior results compared with ConvLSTM, prompting extensive research and development on the ConvGRU and ConvLSTM networks by meteorologists.

Yuan et al. (2018) proposed the model of adaptive ALO-LSTM by combining with ALO module on the basis of LSTM network. It has a better ability to adapt to the data. However, it does not address the problem that the ability of the model to capture long time dependencies still diminishes as the sequence grows. To address this problem, Jing et al. (2020) proposed a hierarchical prediction recurrent neural network (HPRNN) to solve the problem of increasing prediction error over time. MotionGRU (Wu et al. 2021) units were proposed for GRU networks to capture more complex spatiotemporal motions. However, their structures are complex compared with the ConvGRU model. The fusion for iterative networks needs to waste more computational resources.

Additionally, CNNs have been explored for their ability to extract image features efficiently and associate them with label images (Sharif Razavian et al. 2014). A standout model in this category is U-Net (Agrawal et al. 2019), which employs a fully convolutional structure with upsampling and downsampling layers, gaining acclaim for its spatial feature extraction prowess and serving as a foundational network for full CNN models. Nie et al. (2021) proposed SmaAt-UNet on the basis of U-Net network model, which introduces the channel attention module in the model. The feature extraction ability of the model is enhanced. And due to the simplicity of the structure of the U-Net model, meteorologists also used it in multi-source data fusion prediction. The RainNet model, introduced by Ayzel et al. (2020), builds on U-Net by incorporating a recursive network and demonstrates exceptional performance in nowcasting precipitation. Pan et al. (2021) proposed the FURENet model with U-Net as the backbone network, which satisfies the purpose of multivariate information input. Polarized radar parameters specific differential phase (KDP) and differential reflectivity (ZDR) are input into the model to improve the accuracy of precipitation proximity forecasting.

GANs have also been merged with RNN and CNN frameworks to enhance early warning accuracy. Tian et al. (2020) developed the GA-ConvGRU model, incorporating a GAN with the ConvGRU model and outperforming both the ConvGRU and optical flow methods. Wang et al. (2021) proposed an advanced model, ExtGAN, based on a conditional GAN, which surpasses the optical flow method and 3D-CNN models in effectiveness. Notably, the deep generative model of radar (DGMR) by Ravuri et al. (2021) stands out as a seminal contribution, significantly improving the resolution of predicted radar images and consistently outperforming both the U-Net and PySTEPS (Pulkkinen et al. 2019) models in most forecasting scenarios. And due to the wide application of MSE, MAE loss function weighting in traditional methods. It causes blurring of radar images. So many meteorologists use GAN to improve the clarity of radar image. Niu et al. (2024) introduced conditional generative adversarial network (CGAN) network on the basis of encoder-predictor network greatly improved the clarity of the model.

In summary, while the RainNet neural network exhibits certain limitations, conventional models like Rainymotion, based on optical flow (Ayzel et al. 2019), struggle to deliver accurate predictions under heavy precipitation conditions. However, the recursive method utilized within the RainNet model shows promise. In light of this, our study proposes integrating the RainNet network with the ConvGRU network to develop a novel neural network model, the I-ConvGRU (iteration convolutional gated recurrent unit model). This integration leverages the recursive method to shorten the prediction horizon and selects the ConvGRU model for its simplicity to enhance temporal feature extraction within the I-ConvGRU model. This research work offers several advantages over current mainstream methodologies: Firstly, unlike the traditional encoder-predictor, the I-ConvGRU model has a simpler structure, which does not increase the complexity of the model due to the increase of the output signal. In addition, the ConvGRU model is selected as the basic structure, and its spatiotemporal feature extraction ability is obviously higher than that of the convolutional network, and the ConvGRU network structure is relatively simple, saving some computing resources. At the same time, putting the output signal back into the network will introduce more temporal and spatial features, and improve the ability of the network to forecast heavy precipitation. Finally, some jump connections are introduced in the sampling process to help the network introduce multi-scale information. It is helpful to improve the ability of the model to extract features.

The second section of this study details the physical model, experimental data, and evaluation methods employed in deep learning. The third section presents the model's testing results and discussions. The concluding section summarizes the findings of this paper.

Model introduction

Our study introduced the I-ConvGRU model, a novel recurrent neural network that represents a fusion of ConvGRU and RainNet technologies (Figure 1). This model diverges from conventional encoder-predictor frameworks by adopting a unique approach to input and prediction. Specifically, the model forecasts a single subsequent image for each input sequence of images, denoted by j. This predicted image is then recycled back into the input data set, taking the position of the most recent input image. This process is iteratively repeated, allowing for sequential output generation.
Figure 1

I-ConvGRU model structure diagram.

Figure 1

I-ConvGRU model structure diagram.

Close modal

The I-ConvGRU model merges the strengths of traditional encoder-predictor networks with a novel iterative approach. The encoder achieves spatial reduction via downsampling layers, and convolution extracts image spatial features. The decoder then utilizes upsampling layer deconvolution to enlarge image feature dimensions, progressively enhancing resolution. The model is structured with 30 downsampling layers, 33 ConvGRU networks, 30 LeakyReLU activation functions, 33 convolution operations, and 3 upsampling layers. A crucial addition is a 1*1 convolution layer post-tensor splicing aimed at channel compression and feature fusion. The final output stage employs multi-layer convolution to extract spatial features from input, driven by data. The model uses the LeakyReLU activation function prior to linking the downsampling layer with the ConvGRU network, incorporating a minor slope in its negative part to prevent gradient vanishing issues, thereby boosting training stability and effectiveness, as well as enhancing the model's nonlinear fitting capacity (Jiang & Cheng 2019).

This article models based on convolution (ConvGRU), and its basic formula is as follows (Shi et al. 2017):
(1)
(2)
(3)
(4)
For simplicity, the deviation part is omitted from the formula, indicating the convolution operation; indicating the Hadamard product. and represent the memory state, reset gate, update gate, and information, respectively (Figure 2); is the input; f is the activation function, which is ReLU with leakage correction with a negative slope of 0.2. Whenever new input arrives, the reset gate controls whether the previous state is cleared, and the update gate controls the amount of new information entering the state.
Figure 2

The structure of a single GRU neuron in the ConvGRU model.

Figure 2

The structure of a single GRU neuron in the ConvGRU model.

Close modal
Figure 3 outlines the technical methodology of our study. Initially, we randomly selected 20 consecutive images from the HKO-7 data set for training, splitting them into two segments: 10 images served as observed data (true value), and the other 10 served as input data. These input images are processed through the network to predict the radar image for t + 6 min. This predicted image is then incorporated back into the input data set, replacing the oldest image (t − 60), and the process is repeated to sequentially generate predictions up to t + 12 min for 10 predicted images. This iterative recycling of predicted images as new inputs allows for incorporating additional spatiotemporal information, thereby enhancing model optimization. The model parameters of the I-ConvGRU are subsequently refined using the backpropagation algorithm to minimize the loss function, ensuring precise and reliable forecasts.
Figure 3

Technical route for I-ConvGRU model echo extrapolation.

Figure 3

Technical route for I-ConvGRU model echo extrapolation.

Close modal

Test data

In this study, we will use the HKO-7 data set collected by the Hong Kong Hydrological and Meteorological Bureau as data to train and verify the early warning capabilities of the I-ConvGRU model. In order to achieve better prediction results, the data set has undergone denoising preprocessing, and the effectiveness of the data set has been verified on the ConvGRU and TrajGRU models. He et al. (2022) and Zeng et al. (2022) used this data set to verify the model's performance. The HKO-7 data set contains Doppler weather radar observation data in Hong Kong from 2009 to 2015. What is stored is the radar echo CAPPI map of the contour plane after interpolating the reflectivity. The time resolution is 6 min. In order to reduce the dBZ of the image itself, we convert the image reflectivity factor to a grey scale image taking values from 0 to 255.
(5)

This formula represents the pixel value; and represents the reflectivity. The radar echo image is based on the conversion of reflectivity. This formula shows that the larger the pixel value, the greater the reflectivity, the stronger the radar echo intensity, and the greater the rainfall. The images used in the data set are 480*480 resolution images, including a 512*512 km range centered on Hong Kong. To speed up the calculation of the model, we reduce the image size to 256*256 resolution.

We divided the HKO-7 data set into three parts: training set, verification set, and test set. The period from 2009 to 2014 is divided into training and validation sets. The number of days in the training set is 812 days, and the number of days in the validation set is 50 days. The test set selects data from 2015, totaling 131 days. During the model training process, we set the batch size batch_size to 2, the learning rate of Adam to 1 × 10−4, the learning strategy to ReduceLRPlateau, iterations 100,000 times, and the model is saved every 5,000 times, which is consistent with the data of the validation set. Compare and select the best model and then test it. All models are implemented on pytorch and run on NVIDIA Tesia P40 graphics card with 24 GB of video memory.

During the model training process, we convert reflectance into precipitation through the formula:
(6)
where a is equal to 58.53 and b is equal to 1.56. dBZ represents reflectivity, and R represents precipitation. However, due to the uneven rainfall distribution in the HKO-7 data set (Figure 4), the part with rainfall x ≥ 30 mm/h only accounts for 0.42% of the total.
Figure 4

HKO-7 data set rainfall distribution map.

Figure 4

HKO-7 data set rainfall distribution map.

Close modal
To balance the rainfall and improve the prediction rate, we set corresponding weights for each part of the rainfall. This paper selects the weights used in the TrajGRU model.
(7)

Evaluation methods

Common evaluation indicators for regression problems include mean absolute error (MAE) and root mean squared error (MSE), where MAE is the average of the absolute errors between the statistical prediction value and the true value, and MSE is the average sum of the squared errors of the statistical predicted value and the true value. The calculation formulas of MAE and MSE are as follows:
(8)
(9)
In the formula, N is the total number of inputs, represents the true value of the nth picture at the position of the radar echo image (i, j), and represents the size of the predicted value at the same position. Since the rainfall distribution in this paper is uneven, we propose balanced mean absolute error (B-MAE) and balanced mean squared error (B-MSE) as evaluation indicators. Using their sum(1:1) as the loss function (LOSS) improves the prediction performance. The calculation formulas of B-MAE, B-MSE, and LOSS are as follows (Zeng et al. 2022):
(10)
(11)
(12)

is the weight of the nth picture at the position of the radar echo map (i, j). Different weights will be assigned to different positions. The stronger the intensity, the greater the weight to balance the precipitation. This will increase the model's attention to rainfall prediction and better predict heavy rainfall.

Due to different precipitation thresholds, different attention will be paid to meteorological operations. Therefore, evaluating the prediction performance of models under different thresholds is of great significance for echo extrapolation early warning. To assess the prediction performance of the model under different thresholds, we used the critical success index (CSI), false alarm rate (FAR), and Heidke skill score (HSS) for detailed evaluation. The calculation formulas of CSI, FAR, and HSS are as follows:
(13)
(14)
(15)

We treat ground observations as true values, and the result is 1 when the radar echo is higher than the threshold set by the data set and 0 otherwise. In the above formula, True Positive (TP) means (prediction = 1, true value = 1), False Negative (FN) means (prediction = 0, true value = 1), and False Positive (FP) means (prediction = 1, true value = 0) and True Negative (TN) representation (prediction = 0, true value = 0). The higher the CSI and HSS values, the better the prediction effect, and the lower the FAR value, the better the prediction effect.

Meanwhile, in order to study the radar warning pictures in a more refined way, we introduce the structural similarity index (SSIM), and peak signal-to-noise ratio (PSNR), which are indicators to evaluate the quality of the radar warning pictures: the specific formulas of SSIM and PSNR are as follows:
(16)
(17)

The SSIM metrics x and y represent two images, respectively; and represent the luminance averages of the two images; and represent the luminance variance of the two images; is the luminance covariance of the two images; and are the stability parameters. MAX in PSNR denotes the maximum of all values and MSE denotes the root mean square error.

To evaluate the effectiveness of the I-ConvGRU model, we compared it with three other models to assess its performance. The first comparison model is the TrajGRU, an RNN-based model that enhances the GRU network with motion trajectory capabilities, showing improved prediction accuracy over the traditional ConvGRU model. The second model, RainNet, leverages a CNN framework and introduces an iterative network built on the U-Net architecture, allowing for a comparative analysis of early warning capabilities across different models of iterative networks. The third model, Rainymotion, utilizes the optical flow method for precipitation field tracking and adopts a constant vector advection scheme for extrapolation. Rainymotion, part of the rainymotion library, demonstrates performance on par with, or even superior to, the RADVOR (Radar Real-time Forecast) model, covering a broad spectrum of rainfall events. To objectively validate the reliability and generalizability of the I-ConvGRU model, we selected the iteration of each model that yielded the lowest loss function for evaluation on the test set. We employed a comprehensive set of metrics, including MSE, MAE, B-MSE, B-MAE, CSI, HSS, FAR, etc., to conduct a thorough performance analysis of the I-ConvGRU model. The overall performance metrics of MSE, MAE, B-MSE, and B-MAE on the test set are detailed in Figure 5.
Figure 5

Performance of four models on the test set of MSE, MAE, B-MSE, and B-MAE.

Figure 5

Performance of four models on the test set of MSE, MAE, B-MSE, and B-MAE.

Close modal

Data analysis reveals that the I-ConvGRU model outperforms other models in terms of MSE, MAE, B-MSE, and B-MAE for predictions within a 256*256 pixel area, indicating lower discrepancies between predicted images and actual observations. Specifically, compared with the TrajGRU model, which had the next best performance, the I-ConvGRU model's MSE and MAE metrics show reductions of 3.8 and 3.2%, respectively, with B-MSE and B-MAE also decreasing by 5.8 and 3.8%. This proves the superior performance of I-ConvGRU on the test set.

To more intuitively analyze the ability of the test set model to predict pictures. In Table 1, we give the PSNR and SSIM mean values of the pictures predicted by different models on the test set. It can be seen that Rainymotion has the worst structural similarity, followed by the TrajGRU model. The I-ConvGRU model has an SSIM value of 0.7956, which is the closest to 1. Analyzing the PSNRs of the models, it can be seen that the I-ConvGRU model has a PSNR value of 21.77, which is the highest among all the models. This indicates that the I-ConvGRU model has the strongest predictive ability.

Table 1

SSIM and PSNR values of model predictions on the test set for each model

ModelsSSIMPSNR
TrajGRU 0.7874 21.607 
RainNet 0.7899 21.432 
Rainymotion 0.7481 21.076 
I-ConvGRU 0.7956 21.77 
ModelsSSIMPSNR
TrajGRU 0.7874 21.607 
RainNet 0.7899 21.432 
Rainymotion 0.7481 21.076 
I-ConvGRU 0.7956 21.77 

For a more detailed assessment of each model's early warning capabilities, we examined the performance of CSI, FAR, and HSS across different thresholds (as detailed in Tables 24). The effectiveness of CSI and HSS diminishes as the threshold increases. However, the I-ConvGRU model consistently outperforms the comparative models across these metrics, demonstrating greater CSI and HSS values. Notably, at a precipitation threshold of ≥30 mm/h, the I-ConvGRU model's CSI and HSS metrics are increased by 9.6 and 6.8%, respectively, compared with the TrajGRU model. Although its FAR values are less favorable than those of the Rainymotion model at lower precipitation thresholds (R ≥ 2 mm/h, R ≥ 5 mm/h, and R ≥ 10 mm/h), this is likely due to the optical flow method's pixel-level motion analysis, which more accurately captures motion trajectories and reduces false alarm rates. Nonetheless, the I-ConvGRU model shows superior performance across all metrics when the precipitation threshold is set at R ≥ 0.5 mm/h and R ≥ 30 mm/h, which is particularly notable during heavy rainfall events.

Table 2

CSI values of four models at different precipitation thresholds, the best results are in bold

AlgorithmR ≥ 0.5R ≥ 2R ≥ 5R ≥ 10R ≥ 30
TrajGRU 0.6296 0.5672 0.4732 0.3661 0.2611 
RainNet 0.6201 0.5595 0.4635 0.3609 0.2338 
Rainymotion 0.5748 0.5142 0.4197 0.3068 0.1804 
I-ConvGRU 0.6401 0.5794 0.4871 0.3916 0.2862 
AlgorithmR ≥ 0.5R ≥ 2R ≥ 5R ≥ 10R ≥ 30
TrajGRU 0.6296 0.5672 0.4732 0.3661 0.2611 
RainNet 0.6201 0.5595 0.4635 0.3609 0.2338 
Rainymotion 0.5748 0.5142 0.4197 0.3068 0.1804 
I-ConvGRU 0.6401 0.5794 0.4871 0.3916 0.2862 
Table 3

FAR values of four models at different precipitation thresholds, the best results are in bold

AlgorithmR ≥ 0.5R ≥ 2R ≥ 5R ≥ 10R ≥ 30
TrajGRU 0.2514 0.3548 0.4609 0.5541 0.6081 
RainNet 0.2549 0.3601 0.4741 0.5662 0.6213 
Rainymotion 0.2554 0.3153 0.4089 0.5280 0.6806 
I-ConvGRU 0.2437 0.3489 0.4557 0.5424 0.5986 
AlgorithmR ≥ 0.5R ≥ 2R ≥ 5R ≥ 10R ≥ 30
TrajGRU 0.2514 0.3548 0.4609 0.5541 0.6081 
RainNet 0.2549 0.3601 0.4741 0.5662 0.6213 
Rainymotion 0.2554 0.3153 0.4089 0.5280 0.6806 
I-ConvGRU 0.2437 0.3489 0.4557 0.5424 0.5986 
Table 4

HSS values of four models at different precipitation thresholds, the best results are in bold

AlgorithmR ≥ 0.5R ≥ 2R ≥ 5R ≥ 10R ≥ 30
TrajGRU 0.7524 0.7067 0.6274 0.5229 0.4054 
RainNet 0.7436 0.6986 0.6159 0.5152 0.3663 
Rainymotion 0.7049 0.6582 0.5718 0.4498 0.2871 
I-ConvGRU 0.7602 0.7160 0.6390 0.5484 0.4331 
AlgorithmR ≥ 0.5R ≥ 2R ≥ 5R ≥ 10R ≥ 30
TrajGRU 0.7524 0.7067 0.6274 0.5229 0.4054 
RainNet 0.7436 0.6986 0.6159 0.5152 0.3663 
Rainymotion 0.7049 0.6582 0.5718 0.4498 0.2871 
I-ConvGRU 0.7602 0.7160 0.6390 0.5484 0.4331 

Overall, the I-ConvGRU model performs best on MSE, MAE, B-MSE, and B-MAE on the test set. The SSIM and PSNR values predicted by the I-ConvGRU model are also the largest. The pictures that proved their predictions were the most similar to the real ones. At the same time, the model has the best performance compared with other models when the threshold is above 30 mm/h. This is because the traditional method does not introduce the output information, so the prediction performance of the model is not as good as that of the I-ConvGRU method for a small range of heavy rainfall regions. At the same time, the addition of output information provides more features to the model. So its evaluation indicators are in the test set of the best performance.

In our comparative evaluation, we utilized sequences of 10 consecutive radar echoes to forecast the radar echo sequence for the next 10 time steps, effectively predicting 1 h of future data based on the previous hour's historical data. This approach aims to provide a clear and intuitive understanding of the prediction performance across different models. To accurately assess the models' capability to capture the dynamics of severe convective weather, we focused on areas characterized by strong echoes, a common indicator of such weather. Our analysis included cases that span the entire lifecycle of convective processes – namely, the genesis, development, and dissipation stages. This comprehensive approach allows us to evaluate the models' effectiveness in predicting severe convective weather events, highlighting their strengths and limitations in various stages of convective activity.

Case 1

This example selects an extreme short-term precipitation event in Hong Kong between 16:00 and 18:00 on May 23, 2015. The rainfall on that day reached 169.4 mm. As can be seen from Figure 6, we have given the radar echo images of five time periods with prediction times of 6, 18, 30, 42, and 60 min. We use a, b, c, d, and e to represent the ground observation image and TrajGRU, respectively. Rainnet, Rainymotion, and the I-ConvGRU models predict images.
Figure 6

Visualization results of Case 1, (a1–a5) represent the true ground radar values at 6, 18, 30, 42, and 60 min; (b1–b5), (c1–c5), (d1–d5), and (e1–e5) the corresponding time prediction result diagrams of TrajGRU, RainNet, Rainymotion, and I-ConvGRU models, respectively.

Figure 6

Visualization results of Case 1, (a1–a5) represent the true ground radar values at 6, 18, 30, 42, and 60 min; (b1–b5), (c1–c5), (d1–d5), and (e1–e5) the corresponding time prediction result diagrams of TrajGRU, RainNet, Rainymotion, and I-ConvGRU models, respectively.

Close modal

Use the numbers 1, 2, 3, 4, and 5 to represent the echo images at different times. It can be seen from Figure 6 that as the prediction time increases, the prediction ability of the prediction images of each model becomes worse and worse, and the loss of details is greater. There are also certain differences from the ground observation images in the echo size and echo boundary. These problems inevitably appear in the results predicted by the ConvLSTM, CSAConvLSTM (Xiong et al. 2021), and SmaAt-UNet (Nie et al. 2021) models. When the prediction time is 60 min, it can be seen that the I-ConvGRU model performs better. In the strong echo area (black box), it can be seen that although the image of the optical flow method is relatively clear, it overestimates the intensity and range of rainfall, while the convolution-based method, the cumulative RainNet model, underestimates rainfall intensity and extent. However, there is a large deviation between the strong echo position predicted by the TrajGRU model based on the recurrent network and the true position. Only the I-ConvGRU model is basically consistent with the actual echo intensity and echo area.

To analyze the warning ability of the model more intuitively, we present the Taylor diagrams of the different models in Case 1, thus helping to analyze the root mean square deviation, standard deviation, and correlation coefficient between the model simulation results and the observed data. As seen from the plots, the standard deviation of the true value of the 60-min prediction is about 60, indicating that the distribution of precipitation intensity in the radar image is not uniform, thus verifying the effectiveness of the weighted loss function we used for early warning. Additionally, Figure 7 shows that the correlation value of the I-ConvGRU model (red squares) is closest to 1, and the root mean square error is closest to the X-axis observation. Therefore, it can be concluded that the I-ConvGRU model demonstrates the best performance.
Figure 7

Taylor diagrams of radar echograms predicted by different models in Case 1.

Figure 7

Taylor diagrams of radar echograms predicted by different models in Case 1.

Close modal
To specifically quantify the predictive power of the model. We analyzed rainfall's CSI, FAR, and HSS values under three different conditions: precipitation threshold R ≥ 0.5 mm/h, R ≥ 5 mm/h, and R ≥ 30 mm/h. Figure 8 shows the changes in various variables over time under different precipitation thresholds. It can be seen that under different thresholds, as time increases, the CSI and HSS values gradually decrease, and the FAR value gradually increases, corresponding to the test set's results. When the threshold R ≥ 5 mm/h, the FAR value of the model based on Rainymotion is smaller than that of the I-ConvGRU model, which is again verified with the test set results. Analyzing the overall situation, when the threshold value is larger, the HSS, CSI, and FAR values corresponding to the I-ConvGRU model perform better and better than other models. When the threshold R ≥ 30 mm/h is fully analyzed, it is found that the longer the time, the better the effect of the model is compared with other models. The model adds new warning pictures and incorporates new information over time. More strong echo features are extracted, making the model perform better when the threshold is larger. In the case of weak precipitation, the FAR value of the model is worse than that of other models. This may be due to the iterative introduction of more erroneous weak precipitation features, gradually increasing the error. However, when comprehensively comparing the model's effectiveness, its CSI and HSS values perform better than other models. So it can be concluded that the performance of the I-ConvGRU model is better than other models.
Figure 8

In Case 1, the CSI, FAR, and HSS values change over time when the thresholds are 0.5, 5, and 30 mm/h.

Figure 8

In Case 1, the CSI, FAR, and HSS values change over time when the thresholds are 0.5, 5, and 30 mm/h.

Close modal
In order to more effectively determine that the I-ConvGRU model is superior to other models in its warning ability for heavy rainfall events over time.We generated violin plots of radar images for each model to assess early warning performance at 60 min for individual scenarios (Figure 9). Analysis of the Taylor plot indicates that the width of the violin plot for the I-ConvGRU model closely aligns with the actual width, given a reflectivity of 40 dBZ over time. This suggests that the I-ConvGRU model exhibits the most robust warning capability.
Figure 9

Violin plots of reflectivity values of radar echo warning maps for different models at a prediction time of 60 min.

Figure 9

Violin plots of reflectivity values of radar echo warning maps for different models at a prediction time of 60 min.

Close modal

Case 2

In addition, we randomly selected a case with heavy rainfall. The forecast for the time period 13:00–15:00 noon on October 5, 2015, in Hong Kong was selected. The prediction effect is shown in Figure 10. When the prediction time is 60 min, the TrajGRU model and Rainet model in the black box area in the echo map have poor early warning capabilities in heavy precipitation conditions and do not predict the echo boundary and echo intensity similar to the actual map. Although the optical flow method predicts echo intensities similar to ground observation images, there are certain errors in the echo boundaries and echo ranges it predicts. Only the I-ConvGRU model predicts strong echo areas, similar to the actual image echo boundaries. At the same time, it also once again verified the effectiveness of the I-ConvGRU model in predicting strong echo areas compared with other types of models in long-term prediction.
Figure 10

Visualization results of Case 2, (a1–a5) represent the true ground radar values at 6, 18, 30, 42, and 60 min; (b1–b5), (c1–c5), (d1–d5), and (e1–e5) the corresponding time prediction result diagrams of TrajGRU, RainNet, Rainymotion, and I-ConvGRU models, respectively.

Figure 10

Visualization results of Case 2, (a1–a5) represent the true ground radar values at 6, 18, 30, 42, and 60 min; (b1–b5), (c1–c5), (d1–d5), and (e1–e5) the corresponding time prediction result diagrams of TrajGRU, RainNet, Rainymotion, and I-ConvGRU models, respectively.

Close modal
In this example, we evaluated the I-ConvGRU model's performance by calculating the CSI, HSS, and FAR at various precipitation thresholds and intervals, detailed in Figure 11. At lower thresholds (R = 0.5 mm/h and R = 5 mm/h), the I-ConvGRU model's performance closely matched that of the TrajGRU model. However, at a higher threshold (R = 30 mm/h), it exhibited superior CSI and HSS values and a lower FAR than the RainNet and TrajGRU models. These results corroborate previous findings, affirming the I-ConvGRU model's strong predictive capability for heavy rainfall, akin to the performance of the T-Unet model in early warning scenarios of heavy precipitation.
Figure 11

Changes of CSI value, FAR value, and HSS value over time in Case 2 when the thresholds are 0.5, 5, and 30 mm/h.

Figure 11

Changes of CSI value, FAR value, and HSS value over time in Case 2 when the thresholds are 0.5, 5, and 30 mm/h.

Close modal
In addition, we present plots illustrating the temporal evolution of B-MSE and B-MAE values for various models in Figure 12. In-depth verification of the radar echo extrapolation warning process is provided through graphs depicting the temporal evolution of B-MSE and B-MAE values for different models. Analysis reveals that Rainymotion exhibits higher B-MSE and B-MAE values, indicating relatively poorer warning capability. Among the remaining three models, the I-ConvGRU model demonstrates relatively smaller B-MSE and B-MAE values, potentially attributable to its incorporation of output signal information, thereby enhancing early warning capability. Among the remaining three models, the B-MSE and B-MAE values of the I-ConvGRU model are relatively small, which may be due to the fact that the I-ConvGRU model incorporates the information of the output signals and improves the early warning capability.
Figure 12

Plot of B-MAE and B-MSE over time for each model in Case 2.

Figure 12

Plot of B-MAE and B-MSE over time for each model in Case 2.

Close modal

This study innovatively merges the temporal dynamics of the ConvGRU model with the loop iteration strengths of the RainNet model to forge the I-ConvGRU model. This hybrid approach marries the temporal feature extraction capabilities of the ConvGRU network with the iterative optimization advantages of RainNet, enabling a more granular optimization of model parameters through the iterative input of 10 sequential time-series images. This process iteratively forecasts future images, which are then cycled back into the network after removing the initial input image, thus refining the parameters with each iteration. Traditional encoder-predictor networks primarily rely on circular convolutions and fall short of capturing spatial characteristics. To address this, the new model introduces skip connections after each sampling to enrich the model's spatial feature representation.

Comprehensive analysis yields the following conclusions:

  • (1) Visualization of two case studies reveals that the I-ConvGRU model excels in predicting strong echo signals, closely matching the intensity and boundaries of ground-observed echoes. While Rainymotion displays clearer images, its echo boundaries and size predictions significantly deviate from actual observations. This is because the essence of the optical flow method is to predict the movement of pixels. It will not change the size of the pixel value. Similarly, the TrajGRU and RainNet models struggle with accurate echo size predictions. Neither of these two models uses the output signal. Therefore, the echo size feature is insufficient.

  • (2) Comparative testing indicates that the I-ConvGRU model achieves a 3.8 and 3.2% reduction in MSE and MAE, respectively, versus the TrajGRU model; it also sees improvements in B-MSE and B-MAE by 5.8 and 3.8%. Moreover, critical meteorological metrics, CSI and HSS, improve by 9.6 and 6.8% for precipitation rates ≥30 mm/h compared with the TrajGRU model, showcasing superior early warning performance, especially in heavy rainfall scenarios. This is because the traditional encoder-predictor network makes less use of the output signal and cannot extract enough strong echo features.

However, although the I-ConvGRU model improves the accuracy of precipitation proximity forecasting, it, like other deep learning models, faces the situation that it loses too much detail in the later stages of the forecast and is unable to maintain the resolution as in the optical flow method, which leads to a significant decrease in resolution. This is because the widespread use of MSE or MAE loss functions smooths the prediction results, resulting in blurred extrapolated radar images; and the forecasting ability for light rain is weak, which is also reflected in the RainNet model. And there is no guarantee that the best warning rate is maintained in every threshold case. These drawbacks are inevitable in HPRNN (Jing et al. 2020) and ConvLSTM (Shi et al. 2015). Finally, the uneven distribution of HKO-7 data set may affect the early warning result to some extent. In future experiments, in order to increase the resolution, we consider introducing GANs and new loss function boosting models to improve the forecasting ability for light rain and consider using data sets of heavy precipitation to improve the prediction performance of the model. Also consider adding information from multiple input variables to the model for prediction, such as speed and temperature. Humidity and other variables because speed can be more specifically related to time and space. Temperature and humidity because storms, rainfall formation are inextricably linked to temperature. Also, physical constraint equations are considered to optimize the warning rate in combination with deep learning. Physical conditions help the model to better understand the physical meaning of the data. Deep learning helps to learn the nonlinear relationships of the data. Combining the two with each other can better capture the patterns and features in the data and improve the prediction accuracy.

We thank the reviewers for their constructive comments and editorial suggestions that significantly improved the quality of this paper.

This work was sponsored by the National Natural Science Foundation of China (U2342216), Sichuan Provincial Central Leading Local Science and Technology Development Special Project (2023ZYD0147), the Project of the Sichuan Department of Science and Technology (2023NSFSC0244, 2023NSFSC0245), the Open Grants of China Meteorological Administration Radar Meteorology Key Laboratory (2023LRM-A01), and the National Key R&D Program of China (2023YFC3007501).

All relevant data are available from an online repository or repositories. The HKO-7 dataset used in this study is from the Hong Kong Observatory at https://github.com/sxjscien-ce/HKO-7/tree/master/hko_data (accessed on 20 November 2021).

The authors declare there is no conflict.

Agrawal
S.
,
Barrington
L.
,
Bromberg
C.
,
Burge
J.
,
Gazen
C.
&
Hickey
J.
(
2019
)
Machine learning for precipitation nowcasting from radar images. arXiv
.
Ayzel
G.
,
Heistermann
M.
&
Winterrath
T.
(
2019
)
Optical flow models as an open benchmark for radar-based precipitation nowcasting (Rainymotion v0.1)
,
Geoscientific Model Development
,
12
(
4
),
1387
1402
.
http://doi.org/10.5194/gmd-12-1387-2019
.
Ayzel
G.
,
Scheffer
T.
&
Heistermann
M.
(
2020
)
Rainnet v1.0: A convolutional neural network for radar-based precipitation nowcasting
,
Geoscientific Model Development Discussions
,
1
20
.
http://doi.org/10.5194/gmd-2020-30
.
Bauer
P.
,
Thorpe
A.
&
Brunet
G.
(
2015
)
The quiet revolution of numerical weather prediction
,
Nature
,
525
(
7567
),
47
55
.
http://doi.org/10.1038/nature14956
.
Brox
T.
,
Bruhn
A.
,
Papenberg
N.
&
Weickert
J.
(
2004
)
High accuracy optical flow estimation based on a theory for warping
, In
Computer Vision – ECCV 2004
, pp.
25
36
.
Chen
X.
,
Wang
M.
,
Wang
S.
,
Chen
Y.
,
Wang
R.
,
Zhao
C.
&
Hu
X.
(
2022
)
Weather radar nowcasting for extreme precipitation prediction based on the temporal and spatial generative adversarial network
,
Atmosphere
,
13
,
8
.
http://doi.org/10.3390/atmos13081291
.
Chowdhury
A.
,
Egodawatta
P.
&
McGree
J.
(
2023
)
Pattern-based assessment of the influence of rainfall characteristics on urban stormwater quality
,
Water Science & Technology
,
87
(
9
),
2292
2303
.
http://doi.org/10.2166/wst.2023.125
.
Dai
Y.
,
Liu
D.
,
Hu
Q.
&
Yu
X.
(
2022
)
Radar target detection algorithm using convolutional neural network to process graphically expressed range time series signals
,
Sensors
,
22
,
18
.
http://doi.org/10.3390/s22186868
.
Dixon
M.
&
Wiener
G.
(
1993
)
TITAN: Thunderstorm identification, tracking, analysis, and nowcasting – A radar-based methodology
,
Journal of Atmospheric and Oceanic Technology
,
10
(
6
),
785
797
.
https://doi.org/10.1175/1520-0426(1993)010 < 0785:TTITAA > 2.0.CO;2
.
He
W.
,
Xiong
T.
,
Wang
H.
,
He
J.
,
Ren
X.
,
Yan
Y.
&
Tan
L.
(
2022
)
Radar echo spatiotemporal sequence prediction using an improved ConvGRU Deep Learning Model
,
Atmosphere
,
13
,
1
.
http://doi.org/10.3390/atmos13010088
.
Jiang
T.
&
Cheng
J.
(
2019
)
Target recognition based on CNN with LeakyReLU and PReLU activation functions
, In:
2019 International Conference on Sensing, Diagnostics, Prognostics, and Control, SDPC 2019
,
August 15, 2019–August 17, 2019
, pp.
718
722
.
http://doi.org/10.1109/SDPC.2019.00136
.
Jing
J.
,
Li
Q.
,
Peng
X.
,
Ma
Q.
&
Tang
S.
(
2020
)
HPRNN: A hierarchical sequence prediction model for long-term weather radar echo extrapolation
, In:
2020 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2020
,
May 4, 2020–May 8, 2020
, pp.
4142
4146
.
http://doi.org/10.1109/ICASSP40776.2020.9054232
.
Johnson
J. T.
,
MacKeen
P. L.
,
Witt
A.
,
Mitchell
E. D. W.
,
Stumpf
G. J.
,
Eilts
M. D.
&
Thomas
K. W.
(
1998
)
The storm cell identification and tracking algorithm: An enhanced WSR-88D algorithm
,
Weather and Forecasting
,
13
(
2
),
263
276
.
https://doi.org/10.1175/1520-0434(1998)013 < 0263:TSCIAT > 2.0.CO;2
.
Li
L.
,
Schmid
W.
&
Joss
J.
(
1995
)
Nowcasting of motion and growth of precipitation with radar over a complex orography
,
Journal of Applied Meteorology
,
34
(
6
),
1286
1300
.
http://doi.org/10.1175/1520-0450(1995)034 < 1286:NOMAGO > 2.0.CO;2
.
Li
X.
,
Ming
J.
,
Wang
Y.
,
Zhao
K.
&
Xue
M.
(
2013
)
Assimilation of T-TREC-retrieved wind data with WRF 3DVAR for the short-term forecasting of typhoon Meranti (2010) near landfall
,
Journal of Geophysical Research: Atmospheres
,
118
(
18
).
http://doi.org/10.1002/jgrd.50815
.
Lin
C.
,
Vasi
S.
,
Kilambi
A.
,
Turner
B.
&
Zawadzki
I.
(
2005
)
Precipitation forecast skill of numerical weather prediction models and radar nowcasts
,
Geophysical Research Letters
,
32
(
14
),
1
4
.
http://doi.org/10.1029/2005GL023451
.
Min
C.
,
Chen
S.
,
Gourley
J. J.
,
Chen
H.
,
Zhang
A.
,
Huang
Y.
&
Huang
C.
(
2019
)
Coverage of China new generation weather radar network
,
Advances in Meteorology
,
1
10
.
http://doi.org/10.1155/2019/5789358
.
Nie
T.
,
Deng
K.
,
Shao
C.
,
Zhao
C.
,
Ren
K.
&
Song
J.
(
2021
)
Self-attention UNet model for radar based precipitation nowcasting
, In:
6th IEEE International Conference on Data Science in Cyberspace, DSC 2021
,
October 9, 2021–October 11, 2021
, pp.
493
499
.
http://doi.org/10.1109/DSC53577.2021.00079
.
Niu
D.
,
Li
Y.
,
Wang
H.
,
Zang
Z.
,
Jiang
M.
,
Chen
X.
&
Huang
Q.
(
2024
)
FsrGAN: A satellite and radar-based fusion prediction network for precipitation nowcasting
,
IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
,
17
,
7002
7013
.
http://doi.org/10.1109/JSTARS.2024.3376987
.
Pan
X.
,
Lu
Y.
,
Zhao
K.
,
Huang
H.
,
Wang
M.
&
Chen
H.
(
2021
)
Improving nowcasting of convective development by incorporating polarimetric radar variables into a deep-learning model
,
Geophysical Research Letters
,
48
(
21
).
http://doi.org/10.1029/2021GL095302
.
Pulkkinen
S.
,
Nerini
D.
,
Pérez Hortal
A. A.
,
Velasco-Forero
C.
,
Seed
A.
,
Germann
U.
&
Foresti
L.
(
2019
)
Pysteps: An open-source Python library for probabilistic precipitation nowcasting (v1.0)
,
Geoscientific Model Development
,
12
(
10
),
4185
4219
.
http://doi.org/10.5194/gmd-12-4185-2019
.
Ravuri
S.
,
Lenc
K.
,
Willson
M.
,
Kangin
D.
,
Lam
R.
,
Mirowski
P.
,
Fitzsimons
M.
,
Athanassiadou
M.
,
Kashem
S.
,
Madge
S.
,
Prudden
R.
,
Mandhane
A.
,
Clark
A.
,
Brock
A.
,
Simonyan
K.
,
Hadsell
R.
,
Robinson
N.
,
Clancy
E.
,
Arribas
A.
&
Mohamed
S.
(
2021
)
Skilful precipitation nowcasting using deep generative models of radar
,
Nature
,
597
(
7878
),
672
677
.
http://doi.org/10.1038/s41586-021-03854-z
.
Rinehart
R. E.
&
Garvey
E. T.
(
1978
)
Three-dimensional storm motion detection by conventional weather radar
,
Nature
,
273
(
5660
),
287
289
.
http://doi.org/10.1038/273287a0
.
Sharif Razavian
A.
,
Azizpour
H.
,
Sullivan
J.
&
and Carlsson
S.
(
2014
)
CNN features off-the-shelf: An astounding baseline for recognition. arXiv e-prints, arXiv:1403.6382. http://doi.org/10.48550/arXiv.1403.6382
.
Shi
X.
,
Chen
Z.
,
Wang
H.
,
Yeung
D.-Y.
,
Wong
W.-K.
&
Woo
W.-C.
(
2015
)
Convolutional LSTM network: A machine learning approach for precipitation nowcasting
, In:
29th Annual Conference on Neural Information Processing Systems, NIPS 2015
,
December 7, 2015–December 12, 2015
, pp.
802
810
.
Shi
X.
,
Gao
Z.
,
Lausen
L.
,
Wang
H.
,
Yeung
D.-Y.
,
Wong
W.-K.
&
Woo
W.-C.
(
2017
)
Deep learning for precipitation nowcasting: A benchmark and a new model
, In:
31st Annual Conference on Neural Information Processing Systems, NIPS 2017
,
December 4, 2017–December 9, 2017
, pp.
5618
5628
.
Shi
E.
,
Li
Q.
,
Gu
D.
&
Zhao
Z.
(
2018
)
A method of weather radar echo extrapolation based on convolutional neural networks
, In:
24th International Conference on MultiMedia Modeling, MMM 2018
,
February 5, 2018–February 7, 2018
, pp.
16
28
.
http://doi.org/10.1007/978-3-319-73603-7_2
.
Shukla
B. P.
,
Kishtawal
C. M.
&
Pal
P. K.
(
2014
)
Prediction of satellite image sequence for weather nowcasting using cluster-based spatiotemporal regression
,
IEEE Transactions on Geoscience and Remote Sensing
,
52
(
7
),
4155
4160
.
http://doi.org/10.1109/TGRS.2013.2280094
.
Sun
J.
,
Xue
M.
,
Wilson
J. W.
,
Zawadzki
I.
,
Ballard
S. P.
,
Onvlee-Hooimeyer
J.
,
Joe
P.
,
Barker
D. M.
,
Li
P.-W.
,
Golding
B.
,
Xu
M.
&
Pinto
J.
(
2014
)
Use of NWP for nowcasting convective precipitation: Recent progress and challenges
,
Bulletin of the American Meteorological Society
,
95
(
3
),
409
426
.
http://doi.org/10.1175/BAMS-D-11-00263.1
.
Tian
L.
,
Li
X.
,
Ye
Y.
,
Xie
P.
&
Li
Y.
(
2020
)
A generative adversarial gated recurrent unit model for precipitation nowcasting
,
IEEE Geoscience and Remote Sensing Letters
,
17
(
4
),
601
605
.
http://doi.org/10.1109/LGRS.2019.2926776
.
Tishchenko
V. A.
,
Khan
V. M.
,
Kruglova
E. N.
&
Kulikova
I. A.
(
2019
)
Monthly and seasonal prediction of precipitation and air temperature in the Amur River Basin
,
Russian Meteorology and Hydrology
,
44
(
3
),
169
179
.
http://doi.org/10.3103/S1068373919030026
.
Wang
C.
,
Wang
P.
,
Wang
P.
,
Xue
B.
&
Wang
D.
(
2021
)
Using conditional generative adversarial 3-D convolutional neural network for precise radar extrapolation
,
IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
,
14
,
5735
5749
.
http://doi.org/10.1109/JSTARS.2021.3083647
.
Wang
H.
,
Tan
L.
,
Zhang
F.
,
Zheng
J.
,
Liu
Y.
,
Zeng
Q.
,
Yan
Y.
,
Ren
X.
&
Xiang
J.
(
2022
)
Three-dimensional structure analysis and droplet spectrum characteristics of southwest vortex precipitation system based on GPM-DPR
,
Remote Sensing
,
14
,
16
.
http://doi.org/10.3390/rs14164063
.
Wang
H.
,
Li
Z.
,
Zhang
T.
,
Chen
Q.
,
Guo
X.
,
Zeng
Q.
&
Xiang
J.
(
2023
)
Downscaling of GPM satellite precipitation products based on machine learning method in complex terrain and limited observation area
,
Advances in Space Research
,
72
(
6
),
2226
2244
.
http://doi.org/10.1016/j.asr.2023.05.057
.
Woo
W.-C.
&
Wong
W.-K.
(
2017
)
Operational application of optical flow techniques to radar-based rainfall nowcasting
,
Atmosphere
,
8
(
3
).
http://doi.org/10.3390/atmos8030048
.
Wu
H.
,
Yao
Z.
,
Wang
J.
&
Long
M.
(
2021
)
MotionRNN: A flexible model for video prediction with spacetime-varying motions
, In
2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2021
,
June 19, 2021 – June 25, 2021
, pp.
15430
15439
.
http://doi.org/10.1109/CVPR46437.2021.01518
.
Xiang
J.
,
Wang
H.
,
Li
Z.
,
Bu
Z.
,
Yang
R.
&
Liu
Z.
(
2023
)
Case study on the evolution and precipitation characteristics of Southwest Vortex in China: Insights from FY-4A and GPM observations
,
Remote Sensing
,
15
(
16
).
http://doi.org/10.3390/rs15164114
.
Xiong
T.
,
He
J.
,
Wang
H.
,
Tang
X.
,
Shi
Z.
&
Zeng
Q.
(
2021
)
Contextual Sa-Attention convolutional LSTM for precipitation nowcasting: A spatiotemporal sequence forecasting view
,
IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
,
14
,
12479
12491
.
http://doi.org/10.1109/jstars.2021.3128522
.
Yan
Y.
,
Wang
H.
,
Li
G.
,
Xia
J.
,
Ge
F.
,
Zeng
Q.
,
Ren
X.
&
Tan
L.
(
2022
)
Projection of future extreme precipitation in China based on the CMIP6 from a machine learning perspective
,
Remote Sensing
,
14
(
16
).
http://doi.org/10.3390/rs14164033
.
Yin
J.
,
Gao
Z.
&
Han
W.
(
2021
)
Application of a radar echo extrapolation-based deep learning method in strong convection nowcasting
,
Earth and Space Science
,
8
(
8
).
http://doi.org/10.1029/2020ea001621
.
Yuan
X.
,
Chen
C.
,
Lei
X.
,
Yuan
Y.
&
Muhammad Adnan
R.
(
2018
)
Monthly runoff forecasting based on LSTM–ALO model
,
Stochastic Environmental Research and Risk Assessment
,
32
(
8
),
2199
2212
.
http://doi.org/10.1007/s00477-018-1560-y
.
Zeng
Q.
,
Li
H.
,
Zhang
T.
,
He
J.
,
Zhang
F.
,
Wang
H.
,
Qing
Z.
,
Yu
Q.
&
Shen
B.
(
2022
)
Prediction of radar echo space-time sequence based on improving TrajGRU deep-learning model
,
Remote Sensing
,
14
(
19
).
http://doi.org/10.3390/rs14195042
.
Zhang
F.
&
Melhauser
C.
(
2012
)
Practical and intrinsic predictability of severe and convective weather at the mesoscales
,
Journal of the Atmospheric Sciences
,
69
(
11
),
3350
3371
.
http://doi.org/10.1175/jas-d-11-0315.1
.
Zhang
F.
,
Lai
C.
&
Chen
W.
(
2022
)
Weather radar echo extrapolation method based on deep learning
,
Atmosphere
,
13
(
5
).
http://doi.org/10.3390/atmos13050815
.
Zhu
J.
&
Dai
J.
(
2022
)
A rain-type adaptive optical flow method and its application in tropical cyclone rainfall nowcasting
,
Frontiers of Earth Science
,
16
(
2
),
248
264
.
http://doi.org/10.1007/s11707-021-0883-z
.
Zou
H.
,
Wu
S.
,
Shan
J.
&
Yi
X.
(
2019
)
A method of radar echo extrapolation based on TREC and Barnes filter
,
Journal of Atmospheric and Oceanic Technology
,
36
(
9
),
1713
1727
.
http://doi.org/10.1175/JTECH-D-18-0194.1
.
This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (CC BY 4.0), which permits copying, adaptation and redistribution, provided the original work is properly cited (http://creativecommons.org/licenses/by/4.0/).

Supplementary data