The temporal and spatial resolution of rainfall data is crucial for environmental modeling studies in which its variability in space and time is considered as a primary factor. Rainfall products from different remote sensing instruments (e.g., radar, satellite) have different space-time resolutions because of the differences in their sensing capabilities and post-processing methods. In this study, we developed a deep-learning approach that augments rainfall data with increased time resolutions to complement relatively lower-resolution products. We propose a neural network architecture based on Convolutional Neural Networks (CNNs), namely TempNet, to improve the temporal resolution of radar-based rainfall products and compare the proposed model with an optical flow-based interpolation method and CNN-baseline model. While TempNet achieves a mean absolute error of 0.332 mm/h, comparison methods achieve 0.35 and 0.341, respectively. The methodology presented in this study could be used for enhancing rainfall maps with better temporal resolution and imputation of missing frames in sequences of 2D rainfall maps to support hydrological and flood forecasting studies.

  • This paper improves the temporal resolution of radar rainfall products from 10 to 5 min.

  • A residual Convolutional Neural Network architecture, namely TempNet, is proposed for the temporal downscaling of radar rainfall.

  • Results show that the TempNet outperforms the optical flow-based temporal interpolation method in the study area.

From disaster preparedness to response and recovery needs, the availability and quality of environmental datasets have gained significant importance in recent years with the increased impact of natural disasters. Rainfall datasets have been an important component in many modeling applications such as flood forecasting (Seo et al. 2018, 2021; Sit et al. 2021a; Xiang & Demir 2022a), water quality modeling (Jha et al. 2007; Demir et al. 2009), wastewater management (Cahoon & Hanke 2017), along with other data products based on sensor networks and instrumentation (Muste et al. 2017). Since spatial and temporal distributions of rain exert importance in such modeling efforts, the quality and availability of precipitation maps hold the utmost importance in advancing research on disaster mitigation (Teague et al. 2021; Alabbad et al. 2022), risk assessment (Yildirim & Demir 2021), and decision support (Ewing & Demir 2021).

Rainfall products from quantitative precipitation estimation (QPE) systems are three-dimensional, including the time component. The first two dimensions reflect spatial coordinates on earth, including latitude and longitude. The third dimension here is temporal resolution. Weather radars have been a primary instrument for QPE and allow us to capture space-time features of rainfall, which is required for environmental (i.e., hydrologic) predictions at relevant space-time scales. Due to its sensing nature, the accuracy of QPE broadly depends on many factors (Villarini & Krajewski 2010). Since a composite of multiple radars is often used to produce large-scale rain maps, computational capabilities and methodology play an essential role in the process of combining multiple radar data. Consequently, it should be noted that synchronization of different observation times among multiple radars (interpolation or extrapolation over time) is a major challenge to generate a composite rain map. Another QPE issue regarding space-time resolutions of radar product is the misrepresentation of rainfall accumulation arising from radar's intermittent sampling strategy (e.g., Fabry et al. 1994; Liu & Krajewski 1996; Seo & Krajewski 2015).

Mitigating the uncertainty caused by radars in rainfall products has been an essential task in radar hydrology (Krajewski & Smith 2002). Most efforts in radar hydrology for better rainfall datasets rely on understanding uncertainty factors and developing their technical solution in the processing algorithms. Regarding the technology used, efforts usually focus on getting more accurate precipitation data for a better sense of weather in terms of space or time. Once the data have been acquired, however, the effort of generating better datasets becomes a different task.

The temporal resolution of rainfall products is one of the key elements that determine the prediction accuracy of modeling efforts (e.g., Atencia et al. 2011). Interpolating 2D rain maps is not a unique problem for rainfall estimation. The same approach has been a topic of interest in the field of computer vision and, with advancements in computational capacities, in the deep learning applications field, precisely for video frame interpolation, or in other words, video temporal interpolation. Recording every video at high frame rates is impractical or expensive for many settings. Thus, developing models or methods to increase video frame rates has been extensively studied in recent years. In computer vision, temporal interpolation is done with both conventional statistical methods and data-driven machine-learning models. Even though proposed solutions could be realized on 1-channel versions of 2D images that form videos, performing temporal interpolations over 3-channel images is more explored, making the task at hand a problem of estimating the tensor between two 3D tensors. Regardless, the methodology that could be employed and challenges remain vastly similar among video temporal interpolation and temporal interpolation of 2D rain maps.

With inspiration from video frame interpolation studies, this paper proposes a convolutional neural networks (CNNs) based deep learning architecture to account for radar's intermittent sampling and thus improve its product temporal resolution. The proposed CNN model, temporal-resolution network, which will be referred to as TempNet throughout this paper, is compared to three baseline methods: the nearest frame, optical flow, and a relatively simple but deep CNN that will be referred to as CNN-baseline. As with baseline methods to compare, the proposed CNN, TempNet produces an intermediate 2D rain map between two temporally consecutive rain maps. In other words, the methodology presented here aims to perform temporal interpolation of 2D rain maps.

The rest of this paper is structured as follows: in section 2, a brief review of the literature is provided. Subsequently, in section 3, the overview of the methodology is presented, starting with the description of the dataset used, followed by the proposed deep neural networks and baseline methods to compare them. In section 4, the results from the nearest frame, optical flow, CNN-baseline, and the TempNet are presented and discussed. Finally, in section 5, the conclusions are summarized with final remarks regarding the study.

Research studies for better rainfall products can be broadly classified as either quality- or resolution-related. Dataset improvements using deep neural networks include data cleaning to eliminate noise (Lepetit et al. 2021), increasing the resolution or accuracy of datasets by various statistical or data-driven methods (Li et al. 2019; Demiray et al. 2021a, 2021b), synthetic data generation (Gautam et al. 2020), and bias correction (Hu et al. 2021). Resolution-related improvements, on the other hand, either focus on increasing the resolution of two spatial dimensions or the temporal dimension.

Unlike deep learning literature, in meteorology and climatology, whether it is temporal or spatial, super-resolution is referred to as downscaling. Downscaling of precipitation products has been a topic of interest for domain scientists. For instance, Hou et al. (2017) built an ensemble of a Markov chain and a Support Vector Machine to increase the spatial resolution of daily precipitation. Likewise, Vandal et al. (2019) explored five statistical downscaling methods for global climate models (GCMs) including machine learning approaches such as autoencoders and support vector machines. Beyond traditional machine learning models, ANNs have been employed for spatial downscaling as well. Alizamir et al. (2017) integrated particle swarm optimization (PSO) into ANNs for arid regions. Similarly, Salimi et al. (2019) utilized PSO, ANNs, as well as genetic algorithm for downscaling of precipitation events in Austria.

Utilizing adversarial training, Chaudhuri & Robertson (2020) proposed CliGAN, a Generative Adversarial Network (GAN) that uses an encoder–decoder structure for the generator, trained with a combination of adversarial loss, content loss with the Nash–Sutcliff Model Efficiency (NSE), and structural loss with the multi-scale structural similarity index (MSSIM). The authors showed that CliGAN can downscale an ensemble of large-scale annual global maximum precipitation products to regional precipitation products. CNNs were used by Tu et al. (2021). They developed their method using ERA-Interim datasets (European Centre for Medium-Range Weather Forecasts 2006) and then applied it to super-resolution precipitation in Japan's Kuma River Watershed.

Temporal downscaling in the literature is not as profound as spatial downscaling. Lee & Jeong (2014) employed a stochastic take with the genetic algorithm in order to increase the temporal resolution of daily precipitation into hourly. Downscaling daily rain gauge measured precipitation into sub-daily precipitation data with 6 h of temporal resolution, Ryo et al. (2014) used satellite rainfall products. Seo & Krajewski (2015) increased the temporal resolution of rain maps produced by a QPE system from 5 to 1 min using advection correction and validated the results with measurements of ground rain gauges. We are not aware of any other studies that increase the temporal resolution of 2D rainfall maps.

Aside from super-resolution, or downscaling, on the other hand, there are studies that impute missing values in rainfall sequences. In such a study, Norazizi & Deni (2019) utilized ANNs to complete rainfall time series for the data from the Malaysian Meteorological Department. Similarly, for a spatiotemporal imputation of monthly rainfall data in 45 gauge stations located in southwestern Colombia, Canchala-Nastar et al. (2019) trained an ANN. In a similar manner to this study, Gao et al. (2021) tested two deep ANNs for temporal imputation of 2D rainfall maps. They comparatively employed a 3DCNN and a Convolutional Bidirectional LSTM and showed that the LSTM-based method they proposed outperforms the 3DCNN as well as some baselines, namely, optical flow, linear interpolation, and nearest frame.

Since the approach this paper utilizes is similar to video frame interpolation, computer vision literature provides more insight regarding the problem and plausible ways to address it. Besides the widely established method of calculating optical flow with statistical methods in order to generate videos with better frame rates, there are many studies exploring deep neural networks. Niklaus et al. (2017a, 2017b) presented two incremental methods for converting low-temporal-resolution videos to high-temporal-resolution videos that used adaptive convolutions.

In another study, Liu et al. (2017) combined traditional optical flow calculation methods with deep neural networks in a method called deep voxel flow, to both temporally interpolate and extrapolate videos. Beyond that, Jiang et al. (2018) used a neural network structure (Unet) that was built for image segmentation purposes and trained it to learn optical flow to generate slow-motion videos out of regular low frame-rate videos. In a similar but extensive effort, Xiang et al. (2020) proposed a convolutional LSTM-based model for both spatial and temporal interpolation of video frames. Cheng & Chen (2021), using separable convolutions, proposed an approach, namely enhanced deformable separable convolution, that aimed to provide a way to increase the frame rates of videos by estimating more than a frame between two available frames.

Even though the literature over 2D radar rainfall data is quite comprehensive for forecast studies using deep learning, to the best of our knowledge, deep learning has not been explored for the temporal interpolation of 2D radar rainfall prior to this study. Furthermore, temporal interpolation of 2D radar rainfall is not well studied using relatively conventional statistical models. The only method from the literature that had a similar aim as ours employed advection correction, which will be one of the baselines utilized in this study to validate the approach that will be presented in the next section.

This section describes the methods that we have employed in this paper. We start by describing the dataset used, IowaRain (Sit et al. 2021b), in the following subsection. We then provide details about the nearest frame, the optical flow, and the CNN baselines, as well as the CNN-based TempNet neural network architecture proposed in the study, where the baseline and TempNet models will produce a rain map at t = 5, t5 for given consecutive maps at t = 0, t0 and t = 10, t10. Finally, we describe how the neural networks were trained in the study. Process of this study can be seen in Figure 1.
Figure 1

Flowchart summarizing the overall process this study undertakes.

Figure 1

Flowchart summarizing the overall process this study undertakes.

Close modal

Dataset

The dataset used in this paper is a rainfall event dataset, namely, IowaRain. The IowaRain dataset mainly relies on the Iowa Flood Center's QPE system (Seo & Krajewski 2020) using seven NEXRAD radars to cover the entire state of Iowa (Figure 2). The data IowaRain provides 5 min and approximately 500 m resolutions in time and space, respectively, and is actively used in flood forecasting and mapping studies (Hu & Demir 2021; Li et al. 2022). IowaRain consists of 288 rainfall events from 2016 to 2019. Each rainfall event is formed by a set of temporally consecutive 2D rain maps or snapshots for each timestamp with a 5-min interval.
Figure 2

Locations of NEXRAD radars used in IowaRain and the boundary box showing the coverage over and around the state of Iowa.

Figure 2

Locations of NEXRAD radars used in IowaRain and the boundary box showing the coverage over and around the state of Iowa.

Close modal
To prepare the dataset for training and testing, we formed a dataset entry for each snapshot ts in a rain event that has another snapshot coming right after (ts+5) and right before (ts−5) it. Each dataset entry consists of ts−5 and ts+5 as the input and ts as the output (Figure 3). For instance, a sequence of snapshots with a length of 10 would yield eight dataset entries. Then, in order to augment the dataset for better training, we form additional dataset entries by reversing the order of snapshots in the event and, by doing so, we double the number of dataset entries for the training dataset.
Figure 3

An example sequence of 2D rainfall maps and how a set of input and output pairs are built from that sequence.

Figure 3

An example sequence of 2D rainfall maps and how a set of input and output pairs are built from that sequence.

Close modal

The IowaRain dataset chronologically provides 64, 67, 76, and 81 events for each year between 2016 and 2019. In order to do the dataset separation in a fashion closer to a 70/30 split (%70 training, %30 testing), we decided to use the rainfall events for 2019 as the test set and all the rainfall events before 2019 as the training set, making our set lengths 207 and 81 for the training and test sets, respectively. The final dataset entries sum up to 35,258 and 6,725 2D rainfall maps for the training and testing sets, respectively. It should be noted that aside from the test/train split and augmenting the dataset by reversing frame sequences, there was no form of preprocessing. Since the min–max normalization was already applied on the dataset, we did not make any further modifications to the measurements.

Nearest frame

As one of the baselines, we employed the nearest frame as it was previously used in radar rainfall imputation comparisons (Gao et al. 2021). The nearest frame simply refers to assuming the interpolated frame is the same as the temporally closest frame to the forecasted frame. In other words, when interpolating t5 from t0 and t15, t5 is assumed to be the same as t0. In our case of interpolating t5 from t0 and t10, on the other hand, we will use the predecessor frame, t0, even though both input frames have the same distance to the output frame.

Optical flow

In order to form a baseline method that is comparable to the model we propose, we chose to use an optical flow-based temporal interpolation method. This method depends on the optical flow calculation between two 2D rain maps and creating an intermediate frame between those rain maps that is at the same distance temporally. Optical flow has been widely used in the radar rainfall literature (Gao et al. 2021) as a method to build deep learning forecasts upon (Nie et al. 2021; Yan et al. 2021), as a baseline method to compare in the temporal extrapolation of radar rainfall products (Ayzel et al. 2020; Kumar et al. 2020; Xie et al. 2020; Xu et al. 2021) as well as in temporal imputation of 2D rainfall maps (Gao et al. 2021) since it was first proposed by Bowler et al. (2004).

Optical flow is the summary of the spatial changes of objects between two frames of a scene. In other words, it shows the motion of the objects in order to estimate the mobility of velocity fields between two frames. Consequently, an optical flow calculation algorithm is a way to determine how objects move from one scene to another. What an optical flow calculation algorithm actually does is calculate the change of pixel intensities. In computer vision, pixel intensity is defined as the properties a pixel carries in a scene. For a 2D tensor, each value's pixel intensity is determined by its position in the 2D tensor and its value. The same phenomenon is also valid for the case we tackle in this study since rainfall maps are, in fact, 2D tensors.

There are many optical flow calculation algorithms in the computer vision literature. In this paper, we employed the Gunnar–Farneback optical flow (Farnebäck 2003), or sometimes referred to as dense optical flow. Gunnar–Farneback optical flow calculates pixel intensities for each pixel in the scene, as opposed to feature-level intensity calculations, hence the name ‘dense’, as opposed to sparse. For the rain map case, that would mean calculating the changes for each of the measurements for the 0.5 km × 0.5 km areas that make up the 2D rain map. The mathematical formulation of the algorithm is beyond the scope of this paper; we suggest the reader refer to the paper for further details. Once the optical flow is calculated, the next step in the approach is to transfer every measurement depending on their motion vectors in the calculated optical flow and their location in the first and second frames.

For comparison purposes, the optical flow between input snapshots was calculated for each of the entries in the test dataset. Then, color propagation was done using that calculation in order to estimate the intermediate frame. Once all the estimations were done for the testing dataset, the performance metrics reported in the next section were calculated using estimated and actual snapshots. The entirety of the baseline method was implemented using the Gunnar–Farneback optical flow implementation using the OpenCV (Bradski & Kaehler 2000) and NumPy (Harris et al. 2020) numeric computing libraries.

Convolutional neural networks

Convolutional Neural Networks, ConvNets, or simply CNNs work relying on convolutional layers and are primarily used for computer vision tasks. A convolutional layer in a neural network uses weight tensors called filters to explore patterns in inputs. In contrast to weights in a fully connected neural network layer, a filter in a convolutional layer typically works two-dimensionally in order to grasp multi-dimensional (both vertical and horizontal in a 2D black and white image, or in a 2D rain map, similarly) spatial relations over the given input tensor. Building upon this core idea, a CNN with multiple convolutional layers can map complex spatial relations in a given tensor that only one convolutional layer fails to achieve (Goodfellow et al. 2016). For further technical details regarding CNNs, readers could refer to LeCun (1989) and Goodfellow et al. (2016). Considering CNN's applicability and success in multi-dimensional space, where the data comprise spatial representations, and the comprehensive usage of CNNs for 2D rainfall forecasting, CNNs make a sound candidate over soft computing methods for temporal interpolation of radar rainfall, as in radar rainfall datasets, it is crucial to take spatial correlations and patterns into account.

Unlike typical artificial neural networks, CNNs work two-dimensionally over a tensor. Since 2D rainfall maps fall into that description, excluding their temporal dimension, one can assume CNNs are the go-to neural network structures for the interpolation of 2D rainfall maps. Since CNNs are typically used with images containing three color channels, a CNN that processes an input image typically has three filters. The dataset we employed, on the other hand, by its nature, has only one channel, a rain rate (mm/h) map.

To form a CNN-baseline that can be compared to TempNet, we built a four-layer CNN where Rectified Linear Unit (ReLU) was used as the activation function after each layer, namely CNN-baseline. The input frames for the CNN were concatenated into a 3D tensor, as if they were separate channels. In other words, when interpolating t5, t0 of shape 1,088 × 1,760 and t10 of shape 1,088 × 1,760 were concatenated to form the input tensor of shape 2 × 1,088 × 1,760. Details regarding the architecture can be found in Table 1.

Table 1

Architecture details for the CNN-baseline

LayerInput channelsOutput channelsDetails
conv_1 Kernel = 3, stride = 1, padding = 1 
conv_2 Kernel = 3, stride = 1, padding = 1 
conv_3 Kernel = 3, stride = 1, padding = 1 
conv_4 Kernel = 3, stride = 1, padding = 1 
LayerInput channelsOutput channelsDetails
conv_1 Kernel = 3, stride = 1, padding = 1 
conv_2 Kernel = 3, stride = 1, padding = 1 
conv_3 Kernel = 3, stride = 1, padding = 1 
conv_4 Kernel = 3, stride = 1, padding = 1 
Table 2

Contingency table for Hit (H), Miss (M), and False Alarm (F)

Real values
RainNo rain
Predicted values Rain 
No rain – 
Real values
RainNo rain
Predicted values Rain 
No rain – 

TempNet

In order to explore how the temporal resolution of 2D rain maps could be improved, we propose a CNN-based neural network architecture, TempNet. The TempNet provides a fast but effective alternative to the optical flow-based color transfer method and builds upon the CNN architecture we have described in previous subsections.

In a different manner and scale than the CNN-baseline, a CNN that works on a single channel of input was built. However, since the neural network expects two 2D rain maps as input, we had to use two different series of convolutional layers to learn the pattern over them first. Then the difference between those convolutional layers' outputs was added to the first frame in a skip connection fashioned architectural design choice that was introduced to improve the convergence of neural networks over the training data (He et al. 2016). In the end, that summation was fed to another series of convolutional layers to output the intermediate frame between two input frames (Figure 4).
Figure 4

Architecture scheme for the TempNet model.

Figure 4

Architecture scheme for the TempNet model.

Close modal

Both CNN-baseline and the TempNet described here were trained using L1 Loss (also known as mean absolute error – MAE) as the cost function and Adam optimizer (Kingma & Ba 2014) with the help of a Reduce-on-Plateau learning rate scheduler over the training loss. The network was trained on the training dataset described in the Dataset subsection on NVIDIA Titan V GPUs using the PyTorch numeric computing library (Paszke et al. 2019). As for hyperparameters, a batch size of 32, an initial learning rate of 0.001, and a number of epochs of 50 without early stop were used.

This section defines performance metrics and presents results using those metrics for nearest frame, optical flow, CNN-baseline, and the TempNet. The loss changes (L1 Loss) over both training and testing datasets during 50 epochs of training are given in Figures 5 and 6 for the CNN-baseline and the TempNet, respectively. As the figures suggest, the performances of both networks steadily increase over epochs and stabilize after the 40th epoch. It is worth noting that the costs reported in Figures 5 and 6 are averaged over the respective datasets, and they are for normalized datasets.
Figure 5

Change in epoch time, training loss, and test loss over 50 epochs of training for the CNN-baseline.

Figure 5

Change in epoch time, training loss, and test loss over 50 epochs of training for the CNN-baseline.

Close modal
Figure 6

Change in epoch time, training loss, and test loss over 50 epochs of training for the TempNet.

Figure 6

Change in epoch time, training loss, and test loss over 50 epochs of training for the TempNet.

Close modal

We reported six metrics for each of the methods described earlier, namely MAE, root mean square error (RMSE), coefficient of determination (CoD), probability of detection (POD), false alarm ratio (FAR), and critical success index (CSI). MAE reports the mean of the absolute values of the differences between estimated and actual 2D rain maps in the test dataset. RMSE is a similar error metric, but it is calculated by taking the root of the sum of the squared differences between estimated and actual rain maps. On the other hand, CoD is the proportion of dependent variable variation predicted by the independent variable(s), where the best score is 1. We used variance-weighted CoD from the scikit-learn library (Pedregosa et al. 2011) in Python. POD, FAR, and CSI are calculated by (1–3), respectively, using the number of hits (H), false alarms (F), and mises (M). In this case, H, F, and M are calculated in a binary fashion (e.g., threshold is zero) (Table 2); H is defined as the number of correctly estimated rainfall cells, or in other words, the number of elements in the 2D tensor that were correctly estimated as nonzero values. F is defined as the number of wrongly estimated rainfall cells: while the cells were estimated to have rain, the ground truth for the same indices in the 2D map had zero. M is defined as the number of rainfall cells that were estimated as zero, while they have a nonzero value in the ground truth. It should be noted that while the best value for POD and CSI is 1.0 (the greater the better), it is 0.0 for FAR (the lower the better).

Table 3 summarizes the performance of the best CNN-baseline and TempNet models for training over the testing dataset as well as the nearest frame and optical flow's performance using the metrics mentioned above in interpolating t5 from t0 and t10. As Table 3 suggests, both TempNet and CNN-baseline outperform the nearest frame approach and the optical flow-based interpolation baseline method for the MAE metric, over which neural network approaches were trained. In terms of RMSE and CoD, the optical flow appears to be superior, while TempNet seconds outperform the CNN-baseline. It would be fair to assume that if RMSE or CoD were part of the cost function, it might have been trivial for neural network approaches to outperform optical flow over them as well. Among the two neural network models, TempNet provides slightly better results at the expense of a slightly longer training time, as Figures 5 and 6 depict. Thus, for a more accurate temporal-resolution increase, in terms of MAE, while TempNet offers a better solution than optical flow and a basic CNN-baseline, the CNN-baseline also presents good enough results:
(1)
(2)
(3)
Table 3

Performance summary of tested methods for predicting intermediate frame

MethodologyMAE ▾RMSE ▾COD ▴FAR ▾CSI ▴POD▴
Nearest frame 0.533 2.53 0.404 0.102 0.841 0.922 
Optical flow 0.35 1.55 0.729 0.151 0.832 0.975 
CNN-baseline 0.341 1.71 0.699 0.074 0.865 0.928 
TempNet 0.332 1.65 0.72 0.073 0.864 0.925 
MethodologyMAE ▾RMSE ▾COD ▴FAR ▾CSI ▴POD▴
Nearest frame 0.533 2.53 0.404 0.102 0.841 0.922 
Optical flow 0.35 1.55 0.729 0.151 0.832 0.975 
CNN-baseline 0.341 1.71 0.699 0.074 0.865 0.928 
TempNet 0.332 1.65 0.72 0.073 0.864 0.925 

Bold values indicate best scores.

Beyond accuracy, one upside of using neural network models is the runtime. Since calculating optical flow and building the new frame out of two sequential frames needs work over individual pixels or measurements, parallelization is challenging and, consequently, time-consuming. Conversely, the same task takes a trivial amount of time on both GPUs and CPUs compared to the baseline method's runtime using TempNet or CNN-baseline once the training is done. As for POD, the optical flow approach seems to be a better methodology but coupled with the FAR, one can easily suggest that optical flow fills radically more cells with values than other approaches. In other words, optical flow focuses on true positives while estimating significantly more false positives than either of the neural network model estimates. Another take here for neural networks is that optical flow being better at POD is most likely due to the fact that the neural networks we presented were trained by optimizing them over MAE. One can argue that optimizing networks' weights by other metrics would make them better than optical flow as well.

Applying a single-frame temporal interpolation method recursively to generate several intermediate frames is an interesting idea, but there are some drawbacks to this approach. First, recursive temporal interpolation cannot be fully parallelized because certain frames cannot be computed until others are generated. Second, only 2i − 1 intermediate frames can be generated (e.g., 3, 5, 7). Third, during recursive interpolation, errors also build up. Subsequently, in order to understand how recursion affects the performance of the approaches presented, we have run a series of tests. Our first test is to understand how the baseline methods and TempNet perform in interpolating t10 from t0 and t20, namely estimating t10. Table 4 shows each method's performance on this task using the aforementioned metrics. According to the presented scores, the performances of the models change drastically in recursion, and CNN-baseline and TempNet lose first place to optical flow. Furthermore, the change in performance suggests that any of the presented methodologies do not scale well enough. Nevertheless, considering optical flow calculates the difference between two frames and fills in the middle frame by moving the pixels by half of the calculated motion vectors, and TempNet and CNN-baseline performed similarly to optical flow in terms of MAE while providing better FAR and CSI scores, we can infer that neural networks are promising in interpolating the temporal distances they were not trained for. It also deserves mentioning that optical flow seems to spatially over-estimate rainfall again since FAR is significantly greater than that of other approaches.

Table 4

Performance summary of methodologies in interpolating frames 10-min apart

MethodologyMAE ▾RMSE ▾COD ▴FAR ▾CSI ▴POD▴
Nearest frame 0.728 3.24 0.003 0.152 0.739 0.846 
Optical flow 0.499 2.25 0.504 0.208 0.768 0.961 
CNN-baseline 0.506 2.40 0.451 0.136 0.788 0.896 
TempNet 0.503 2.38 0.458 0.129 0.788 0.888 
MethodologyMAE ▾RMSE ▾COD ▴FAR ▾CSI ▴POD▴
Nearest frame 0.728 3.24 0.003 0.152 0.739 0.846 
Optical flow 0.499 2.25 0.504 0.208 0.768 0.961 
CNN-baseline 0.506 2.40 0.451 0.136 0.788 0.896 
TempNet 0.503 2.38 0.458 0.129 0.788 0.888 

Bold values indicate best scores.

Furthermore, to see how errors build up in recursion, we performed a test where an already interpolated frame of t10 from t0 and t20 was used. In other words, (1) frame t10 was interpolated using t0 and t20, and (2) using the newly interpolated frame t10, and ground truth frames of t0 and t20, t5 and t15 were interpolated, respectively (Figure 7). Table 5 shows the performance of each of the proposed methods in iterative estimation. While neural network models present the best results in terms of the average absolute error, they are still not as good as the optical flow approach in providing a solution when it comes to POD. Nevertheless, the fact that optical flow is drastically worse at FAR changes the tie in favor of neural network-based methods.
Table 5

Performances of methodologies in interpolating frames in the second iteration

MethodologyMAE ▾RMSE ▾COD ▴FAR ▾CSI ▴POD▴
Nearest frame 0.686 3.05 0.080 0.144 0.753 0.854 
Optical flow 0.593 2.58 0.361 0.215 0.757 0.952 
CNN-baseline 0.422 2.05 0.585 0.088 0.820 0.888 
TempNet 0.422 2.06 0.586 0.089 0.819 0.888 
MethodologyMAE ▾RMSE ▾COD ▴FAR ▾CSI ▴POD▴
Nearest frame 0.686 3.05 0.080 0.144 0.753 0.854 
Optical flow 0.593 2.58 0.361 0.215 0.757 0.952 
CNN-baseline 0.422 2.05 0.585 0.088 0.820 0.888 
TempNet 0.422 2.06 0.586 0.089 0.819 0.888 

Bold values indicate best scores.

Figure 7

Workflow of iterative estimation of t5 and t15.

Figure 7

Workflow of iterative estimation of t5 and t15.

Close modal
Figures 8 and 9 show ground truth and each methodology's output in interpolating frames 5 min apart by using the color scheme of the Iowa Flood Information System (Demir et al. 2018). According to the rainfall maps depicted in these figures, we can easily see that radar rainfall maps have small areas of precipitation scattered around larger areas that are denser. This phenomenon could be because of the artifacts caused by radars or the QPE process. Optical flow seems to be transferring those artifact-like small regions relatively better than other approaches. However, it also distributes the artifacts as well as larger regions and forms shadows of small precipitation regions all around the map. Coupled with Table 3, optical flow's ability to transfer small areas of precipitation makes them in-overall better at POD, and creating those shadows makes them estimate false positives, thus resulting in a worse FAR score. On the other hand, performing similarly, CNN-baseline and TempNet get rid of those artifact-like small regions by producing relatively blurry outputs. This tendency to ignore small areas to focus on larger areas comes with an expense. More often than not, intense rainfall happens in smaller areas, and since the CNN-based approaches we presented tend to ignore changes in small areas, they are also inclined to not estimate intense rainfall as accurately as optical flow.
Figure 8

An example 10-min sequence of ground truth and estimations by optical flow, CNN-baseline, and TempNet from a rainfall event on 8 May 2019 (zoomed for visualization).

Figure 8

An example 10-min sequence of ground truth and estimations by optical flow, CNN-baseline, and TempNet from a rainfall event on 8 May 2019 (zoomed for visualization).

Close modal
Figure 9

An example 10-min sequence of ground truth and estimations by optical flow, CNN-baseline, and TempNet from a rainfall event on 11 June 2019 (zoomed for visualization).

Figure 9

An example 10-min sequence of ground truth and estimations by optical flow, CNN-baseline, and TempNet from a rainfall event on 11 June 2019 (zoomed for visualization).

Close modal

Finally, for comparison purposes, we did some tests to understand the visual performance of the CNN-baseline and the TempNet by increasing the temporal resolution of some events in the test dataset by three iterations over the resolutions they were not trained for. In other words, rainfall maps' temporal resolution was increased from 5 to 0.7 min. After visualizing the actual and generated rain maps with the Iowa Flood Information System's rainfall map color scheme, even though we were able to see that the TempNet and the CNN-baseline were not able to carry all the information regarding the rainfall in specific areas, they were still able to represent most of the accumulated rainfall clearly. Alternatively, the optical flow method would visually average two input frames that had lots of shadows scattered over the maps.

In summary, it was demonstrated that neural network-based approaches outperform optical flow-based advection correction in the literature for the metric that they were trained for. This is largely due to the fact that the neural networks presented here are designed to explore nonlinear patterns as well as linear ones, while optical flow calculations work purely on linear connections. The difference between TempNet and CNN-baseline, on the other hand, is thanks to the residual connection, where the learned patterns and the original data are combined for better estimations. In order for TempNet or CNN-baseline to be more competitive over RMSE and CoD, one could easily incorporate them into the cost function for better scores. As for rainfall metrics such as CSI and POD, since they do not comprise formulas that the derivatives could easily be taken from, they cannot be part of the cost function right off the bat.

In this study, we presented a CNN-based temporal-resolution improvement model, TempNet, and compared it to the nearest frame approach, an optical flow-based baseline method, and a CNN-based baseline method, namely CNN-baseline. Results show that TempNet broadly outperforms baseline models according to MAE as well as a combination of other metrics such as FAR, POD, and CSI, while doing that with a significantly better computation time than optical flow. We consider this work as a significant step towards creating better rainfall maps for hydrological modeling needs such as flood forecasting (Sit et al. 2021a; Xiang et al. 2021), climate change modeling (Rolnick et al. 2019; Sit et al. 2020), and missing rainfall map imputation (Gao et al. 2021). Nevertheless, it comes with shortcomings. One shortcoming of TempNet that strikes first is the fact that it does not perform as well as optical flow in terms of some metrics such as RMSE and CoD. Even though, as previously stated, this is due to the cost function selection, a better neural network that performs better across those metrics is theoretically possible. However, since TempNet is a relatively small neural network with fast training and estimation times, this tendency is somewhat expected. Making TempNet more complex at the expense of computation time could be the solution.

This study and the approaches provided here were only tested over a dataset that covered the state of Iowa; thus, one of the shortcomings here is that the generalization performance of the approaches is not known. As a future direction, we aim to develop new iterations of TempNet and CNN-baseline and test them over datasets that cover wider areas, potentially the whole earth. Also, to support operational flood forecasting needs (Xiang & Demir 2022b), we aim to extend the neural networks presented in this study in two different directions: (1) deeper neural networks for better interpolation accuracy; and (2) neural networks with similar performance but doing so with faster computation times by building more efficient neural networks. In order to train better models, various loss functions could be utilized, such as a loss function where the metrics we used in this are combined by differentiable weights.

One challenge with the process was that weather radar-based rainfall maps contain significant artifacts. Although the IowaRain dataset provides rain maps that are vastly cleansed of those artifacts, some remaining noise (e.g., ground and wind turbine clutter; Seo et al. 2015) may affect both accuracy and visual representations. In order to test the dataset better and to develop more accurate models for both imputation and temporal super-resolution of radar rainfall products, datasets with fewer artifacts would become useful. Thus, in order to create improved datasets, as a future perspective, we aim to focus on denoising radar rainfall maps to decrease the number of artifacts in the original dataset with neural networks and train the TempNet over a cleaner dataset. This highlights the need for better benchmark datasets for research communities (Demir et al. 2022). Another future aspect of this problem would be to rectify the issues we mentioned regarding recursive temporal interpolation by introducing a neural network that is capable of producing intermediate frames for any given temporal distance to the input frames, as in Jiang et al. (2018).

All relevant data are included in the paper or its Supplementary Information.

The authors declare there is no conflict.

Alizamir
M.
,
Azhdary Moghadam
M.
,
Hashemi Monfared
A.
&
Shamsipour
A.
2017
A hybrid artificial neural network and particle swarm optimization algorithm for statistical downscaling of precipitation in arid region
.
Ecopersia
5
(
4
),
1991
2006
.
Atencia
A.
,
Mediero
L.
,
Llasat
M. C.
&
Garrote
L.
2011
Effect of radar rainfall time resolution on the predictive capability of a distributed hydrologic model
.
Hydrology and Earth System Sciences
15
(
12
),
3809
3827
.
Ayzel
G.
,
Scheffer
T.
&
Heistermann
M.
2020
Rainnet v1. 0: a convolutional neural network for radar-based precipitation nowcasting
.
Geoscientific Model Development
13
(
6
),
2631
2644
.
Bowler
N. E.
,
Pierce
C. E.
&
Seed
A.
2004
Development of a precipitation nowcasting algorithm based upon optical flow techniques
.
Journal of Hydrology
288
(
1–2
),
74
91
.
Bradski
G.
&
Kaehler
A.
2000
OpenCV
.
Dr. Dobb's Journal of Software Tools
25
,
120
123
.
Canchala-Nastar
T.
,
Carvajal-Escobar
Y.
,
Alfonso-Morales
W.
,
Cerón
W. L.
&
Caicedo
E.
2019
Estimation of missing data of monthly rainfall in southwestern Colombia using artificial neural networks
.
Data in Brief
26
,
104517
.
Cheng
X.
&
Chen
Z.
2021
Multiple video frame interpolation via enhanced deformable separable convolution
.
IEEE Transactions on Pattern Analysis and Machine Intelligence.
44
(
10
),
7029
7045
.
DOI: http://doi.org/10.1109/TPAMI.2021.3100714
.
Demir
I.
,
Jiang
F.
,
Walker
R. V.
,
Parker
A. K.
&
Beck
M. B.
2009
Information systems and social legitimacy scientific visualization of water quality
. In
2009 IEEE International Conference on Systems, Man and Cybernetics
.
IEEE
, pp.
1067
1072
.
Demir
I.
,
Yildirim
E.
,
Sermet
Y.
&
Sit
M. A.
2018
FLOODSS: Iowa flood information system as a generalized flood cyberinfrastructure
.
International Journal of River Basin Management
16
(
3
),
393
400
.
Demir
I.
,
Xiang
Z.
,
Demiray
B.
&
Sit
M.
2022
Waterbench: a large-scale benchmark dataset for data-driven streamflow forecasting
.
Earth System Science Data
14
(
12
),
5605
5616
.
DOI: https://doi.org/10.5194/essd-14-5605-2022
.
Demiray
B. Z.
,
Sit
M.
&
Demir
I.
2021a
D-SRGAN: DEM super-resolution with generative adversarial networks
.
SN Computer Science
2
(
1
),
1
11
.
Demiray
B. Z.
,
Sit
M.
&
Demir
I.
2021b
DEM Super-Resolution with EfficientNetV2. arXiv preprint arXiv:2109.09661
.
ERA-Interim, European Centre for Medium-Range Weather Forecasts
2006
Fabry
F.
,
Bellon
A.
,
Duncan
M. R.
&
Austin
G. L.
1994
High resolution rainfall measurements by radar for very small basins: the sampling problem reexamined
.
Journal of Hydrology
161
(
1–4
),
415
428
.
Farnebäck
G.
2003
Two-frame motion estimation based on polynomial expansion
. In:
Image Analysis: 13th Scandinavian Conference, SCIA 2003
.
Halmstad, Sweden, June 29–July 2, Proceedings 13. Springer, Berlin Heidelberg, pp. 363–370. https://doi.org/10.1007/3-540-45103-X_50
.
Gao
L.
,
Zheng
Y.
,
Wang
Y.
,
Xia
J.
,
Chen
X.
,
Li
B.
,
Luo
M.
&
Guo
Y.
2021
Reconstruction of missing data in weather radar image sequences using deep neuron networks
.
Applied Sciences
11
(
4
),
1491
.
Gautam
A.
,
Sit
M.
&
Demir
I.
2020
Realistic river image synthesis using deep generative adversarial networks
.
Front. Water
4
,
784441
.
doi:10.3389/frwa.2022.784441
.
Goodfellow
I.
,
Bengio
Y.
&
Courville
A.
2016
Deep Learning
.
MIT Press
,
Cambridge, MA, USA
.
Harris
C. R.
,
Millman
K. J.
,
van der Walt
S. J.
,
Gommers
R.
,
Virtanen
P.
,
Cournapeau
D.
,
Wieser
E.
,
Taylor
J.
,
Berg
S.
,
Smith
N. J.
&
Kern
R.
2020
Array programming with NumPy
.
Nature
585
(
7825
),
357
362
.
He
K.
,
Zhang
X.
,
Ren
S.
&
Sun
J.
2016
Deep residual learning for image recognition
. In:
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
, pp.
770
778
.
Hou
Y. K.
,
Chen
H.
,
Xu
C. Y.
,
Chen
J.
&
Guo
S. L.
2017
Coupling a Markov chain and support vector machine for at-site downscaling of daily precipitation
.
Journal of Hydrometeorology
18
(
9
),
2385
2406
.
Hu
Y. F.
,
Yin
F. K.
&
Zhang
W. M.
2021
Deep learning-based precipitation bias correction approach for Yin–He global spectral model
.
Meteorological Applications
28
(
5
),
e2032
.
Jha
M. K.
,
Gassman
P. W.
&
Arnold
J. G.
2007
Water quality modeling for the Raccoon River watershed using SWAT
.
Transactions of the ASABE
50
(
2
),
479
493
.
Jiang
H.
,
Sun
D.
,
Jampani
V.
,
Yang
M. H.
,
Learned-Miller
E.
&
Kautz
J.
2018
Super slomo: High quality estimation of multiple intermediate frames for video interpolation
. In:
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
, pp.
9000
9008
.
Kingma
D. P.
&
Ba
J.
2014
Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980
.
Krajewski
W. F.
&
Smith
J. A.
2002
Radar hydrology: rainfall estimation
.
Advances in Water Resources
25
(
8–12
),
1387
1394
.
LeCun
Y.
1989
Generalization and network design strategies
.
Connectionism in Perspective
19
,
143
155
.
Lepetit
P.
,
Ly
C.
,
Barthès
L.
,
Mallet
C.
,
Viltard
N.
,
Lemaitre
Y.
&
Rottner
L.
2021
Using deep learning for restoration of precipitation echoes in radar data
.
IEEE Transactions on Geoscience and Remote Sensing
60
,
1
14
.
Li
W.
,
Zhou
W.
,
Wang
Y. M.
,
Shen
C.
,
Zhang
X.
&
Li
X.
2019
Meteorological radar fault diagnosis based on deep learning
. In:
2019 International Conference on Meteorology Observations (ICMO)
.
IEEE
,
Chengdu, China
, pp.
1
4
.
Liu
C.
&
Krajewski
W. F.
1996
A comparison of methods for calculation of radar-rainfall hourly accumulations 1
.
JAWRA Journal of the American Water Resources Association
32
(
2
),
305
315
.
Liu
Z.
,
Yeh
R. A.
,
Tang
X.
,
Liu
Y.
&
Agarwala
A.
2017
Video frame synthesis using deep voxel flow
. In:
Proceedings of the IEEE International Conference on Computer Vision
, pp.
4463
4471
.
DOI: http://doi.org/10.1109/ICCV.2017.478
.
Muste
M.
,
Lyn
D. A.
,
Admiraal
D.
,
Ettema
R.
,
Nikora
V.
&
García
M. H.
2017
Experimental Hydraulics: Methods, Instrumentation, Data Processing and Management: Volume I: Fundamentals and Methods
.
CRC Press
,
London, UK
.
Nie
T.
,
Ji
X.
&
Pang
Y.
2021
OFAF-ConvLSTM: an optical flow attention fusion-ConvLSTM model for precipitation nowcasting
. In:
2021 3rd International Academic Exchange Conference on Science and Technology Innovation (IAECST)
.
IEEE
,
Guangzhou, China
, pp.
283
286
.
Niklaus
S.
,
Mai
L.
&
Liu
F.
2017a
Video frame interpolation via adaptive convolution
. In:
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
, pp.
670
679
.
Niklaus
S.
,
Mai
L.
&
Liu
F.
2017b
Video frame interpolation via adaptive separable convolution
. In:
Proceedings of the IEEE International Conference on Computer Vision
, pp.
261
270
.
Norazizi
N. A. A.
&
Deni
S. M.
2019
Comparison of Artificial Neural Network (ANN) and Other Imputation Methods in Estimating Missing Rainfall Data at Kuantan Station
. In:
Berry, M., Yap, B., Mohamed, A., Köppen, M. (eds). Soft Computing in Data Science. SCDS 2019. Communications in Computer and Information Science, 1100. Springer, Singapore. https://doi.org/10.1007/978-981-15-0399-3_24
.
Paszke
A.
,
Gross
S.
,
Massa
F.
,
Lerer
A.
,
Bradbury
J.
,
Chanan
G.
,
Killeen
T.
,
Lin
Z.
,
Gimelshein
N.
,
Antiga
L.
&
Desmaison
A.
2019
Pytorch: an imperative style, high-performance deep learning library
.
Advances in Neural Information Processing Systems
32
,
8026
8037
.
Pedregosa
F.
,
Varoquaux
G.
,
Gramfort
A.
,
Michel
V.
,
Thirion
B.
,
Grisel
O.
,
Blondel
M.
,
Prettenhofer
P.
,
Weiss
R.
,
Dubourg
V.
&
Vanderplas
J.
2011
Scikit-learn: Machine learning in Python
.
The Journal of Machine Learning Research
12
,
2825
2830
.
Rolnick
D.
,
Donti
P. L.
,
Kaack
L. H.
,
Kochanski
K.
,
Lacoste
A.
,
Sankaran
K.
,
Ross
A. S.
,
Milojevic-Dupont
N.
,
Jaques
N.
,
Waldman-Brown
A.
&
Luccioni
A.
2019
Tackling climate change with machine learning. arXiv preprint arXiv:1906.05433
.
Salimi
A. H.
,
Masoompour Samakosh
J.
,
Sharifi
E.
,
Hassanvand
M. R.
,
Noori
A.
&
von Rautenkranz
H.
2019
Optimized artificial neural networks-based methods for statistical downscaling of gridded precipitation data
.
Water
11
(
8
),
1653
.
Seo
B. C.
,
Krajewski
W. F.
,
Quintero
F.
,
Buan
S.
&
Connelly
B.
2021
Assessment of streamflow predictions generated using multimodel and multiprecipitation product forcing
.
Journal of Hydrometeorology
22
(
9
),
2275
2290
.
Sit
M.
,
Demiray
B. Z.
,
Xiang
Z.
,
Ewing
G. J.
,
Sermet
Y.
&
Demir
I.
2020
A comprehensive review of deep learning applications in hydrology and water resources
.
Water Science and Technology
82
(
12
),
2635
2670
.
Sit
M.
,
Demiray
B.
&
Demir
I.
2021a
Short-term hourly streamflow prediction with graph convolutional GRU networks. arXiv preprint arXiv:2107.07039
.
Sit
M.
,
Seo
B. C.
&
Demir
I.
2021b
IowaRain: a statewide rain event dataset based on weather radars and quantitative precipitation estimation. arXiv preprint arXiv:2107.03432
.
Teague
A.
,
Sermet
Y.
,
Demir
I.
&
Muste
M.
2021
A collaborative serious game for water resources planning and hazard mitigation
.
International Journal of Disaster Risk Reduction
53
,
101977
.
Tu
T.
,
Ishida
K.
,
Ercan
A.
,
Kiyama
M.
,
Amagasaki
M.
&
Zhao
T.
2021
Hybrid precipitation downscaling over coastal watersheds in Japan using WRF and CNN
.
Journal of Hydrology: Regional Studies
37
,
100921
.
Vandal
T.
,
Kodra
E.
&
Ganguly
A. R.
2019
Intercomparison of machine learning methods for statistical downscaling: The case of daily and extreme precipitation
.
Theoretical and Applied Climatology
137
,
557
570
.
Xiang
Z.
&
Demir
I.
2022a
Fully distributed rainfall-runoff modeling using spatial-temporal graph neural network
.
EarthArxiv
3018
.
doi:10.31223/X57P74
.
Xiang
Z.
&
Demir
I.
2022b
Real-Time streamflow forecasting framework, implementation and post-analysis using deep learning
.
EarthArxiv
3162
.
https://doi.org/10.31223/X5BW6R
Xiang
X.
,
Tian
Y.
,
Zhang
Y.
,
Fu
Y.
,
Allebach
J. P.
&
Xu
C.
2020
Zooming slow-mo: Fast and accurate one-stage space-time video super-resolution
. In:
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
, pp.
3370
3379
.
Xiang
Z.
,
Demir
I.
,
Mantilla
R.
&
Krajewski
W. F.
2021
A regional semi-distributed streamflow model using deep learning
.
EarthArxiv
2152
.
doi:10.31223/X5GW3V
.
Xie
P.
,
Li
X.
,
Ji
X.
,
Chen
X.
,
Chen
Y.
,
Liu
J.
&
Ye
Y.
2020
An energy-based generative adversarial forecaster for radar echo map extrapolation
.
IEEE Geoscience and Remote Sensing Letters
19
,
1
5
.
DOI: https://doi.org/10.1109/LGRS.2020.3023950
.
Xu
F.
,
Li
G.
,
Du
Y.
,
Chen
Z.
&
Lu
Y.
2021
Multi-layer networks for ensemble precipitation forecasts postprocessing
. In:
Proceedings of the AAAI Conference on Artificial Intelligence
,
35
(
17
), pp.
14966
14973
.
DOI: https://doi.org/10.1609/aaai.v35i17.17756
.
Yan
B. Y.
,
Yang
C.
,
Chen
F.
,
Takeda
K.
&
Wang
C.
2021
FDNet: a deep learning approach with two parallel cross encoding pathways for precipitation nowcasting. arXiv preprint arXiv:2105.02585
.
This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (CC BY 4.0), which permits copying, adaptation and redistribution, provided the original work is properly cited (http://creativecommons.org/licenses/by/4.0/).