Articles | Volume 15, issue 15
https://doi.org/10.5194/gmd-15-5967-2022
https://doi.org/10.5194/gmd-15-5967-2022
Model description paper
 | 
01 Aug 2022
Model description paper |  | 01 Aug 2022

Rad-cGAN v1.0: Radar-based precipitation nowcasting model with conditional generative adversarial networks for multiple dam domains

Suyeon Choi and Yeonjoo Kim
Abstract

Numerical weather prediction models and probabilistic extrapolation methods using radar images have been widely used for precipitation nowcasting. Recently, machine-learning-based precipitation nowcasting models have also been actively developed for relatively short-term precipitation predictions. This study was aimed at developing a radar-based precipitation nowcasting model using an advanced machine-learning technique, conditional generative adversarial network (cGAN), which shows high performance in image generation tasks. The cGAN-based precipitation nowcasting model, named Rad-cGAN, developed in this study was trained with the radar reflectivity data of the Soyang-gang Dam basin in South Korea with a spatial domain of 128 × 128 pixels, spatial resolution of 1 km, and temporal resolution of 10 min. The model performance was evaluated using previously developed machine-learning-based precipitation nowcasting models, namely convolutional long short-term memory (ConvLSTM) and U-Net. In addition, Eulerian persistence model and pySTEPS, a radar-based deterministic nowcasting system, are used as baseline models.

We demonstrated that Rad-cGAN outperformed reference models at 10 min lead time prediction for the Soyang-gang Dam basin based on verification metrics: Pearson correlation coefficient (R), root mean square error (RMSE), Nash–Sutcliffe efficiency (NSE), critical success index (CSI), and fraction skill scores (FSS) at an intensity threshold of 0.1, 1.0, and 5.0 mm h−1. However, unlike low rainfall intensity, the CSI at high rainfall intensity in Rad-cGAN deteriorated rapidly beyond the lead time of 10 min; however, ConvLSTM and baseline models maintained better performances. This observation was consistent with the FSS calculated at high rainfall intensity. These results were qualitatively evaluated using typhoon Soulik as an example, and through this, ConvLSTM maintained relatively higher precipitation than the other models. However, for the prediction of precipitation area, Rad-cGAN showed the best results, and the advantage of the cGAN method to reduce the blurring effect was confirmed through radially averaged power spectral density (PSD). We also demonstrated the successful implementation of the transfer learning technique to efficiently train the model with the data from other dam basins in South Korea, such as the Andong Dam and Chungju Dam basins. We used the pre-trained model, which was completely trained in the Soyang-gang Dam basin. Furthermore, we analyzed the amount of data to effectively develop the model for the new domain through the transfer learning strategies applying the pre-trained model using data for additional dam basins. This study confirmed that Rad-cGAN can be successfully applied to precipitation nowcasting with longer lead times and using the transfer learning approach showed good performance in dam basins other than the originally trained basin.

Dates
1 Introduction

Nowcasting is defined as a description of the current weather and then forecasting within a few hours and is generally applied to mesoscale and local scales. Owing to the increasing number of disasters on small spatiotemporal scales, nowcasting plays an important role in risk management (Wang et al., 2017). Therefore, the need for accurate precipitation nowcasting for early warning systems is increasing to reduce the damage caused by heavy rain, landslides, and flash floods.

Among the existing precipitation nowcasting models, numerical weather prediction (NWP), which performs rainfall prediction based on atmospheric physics equations, can generate high-resolution rainfall forecasts with long lead times. However, NWP has exhibited poor forecast performance with relatively short (0–2 h) lead times (Berenguer et al., 2012). Several studies have demonstrated that radar-based models based on the extrapolation method perform better than NWP, especially in the case of precipitation nowcasting with lead times of up to 6 h (Berenguer et al., 2012; Pierce et al., 2012; Renzullo et al., 2017; Imhoff et al., 2020). Additionally, the increased availability of high-resolution remote sensing observation data (e.g., radar) and computer resources has facilitated the development of advanced precipitation nowcasting models. For example, Ayzel et al. (2019) developed an optical flow-based precipitation nowcasting model called rainymotion, and Pulkkinen et al. (2019) developed a deterministic and probabilistic nowcasting application called pySTEPS, which has potential applications in several countries (Finland, Switzerland, the United States, and Australia). Both models were written in an open-source Python library. Furthermore, the blending technique, which combines NWP and radar-based models, has improved the precipitation nowcasting performance for short-term flood forecasting (Poletti et al., 2019; Hwang et al., 2020).

Recent availability of a large amount of data and increased computational resources led to the development of radar-based models using machine-learning techniques. Shi et al. (2015) developed a radar-based model with a convolutional long short-term memory (ConvLSTM) architecture that outperformed the optical flow-based model. They showed that ConvLSTM can capture the spatiotemporal correlation between input rainfall image frames, which are recorded every 6 min across Hong Kong. In addition, several studies have shown that the ConvLSTM architecture can be successfully applied to the precipitation nowcasting model (Kim et al., 2017; Moishin et al., 2021; Sønderby et al., 2020; Jeong et al., 2021). Although the convolution neural network (CNN) does not have a structure to conserve temporal information, Agrawal et al. (2019) showed that a fully connected CNN called U-Net can make better predictions than traditional NWP models. Further studies (e.g., Ayzel et al., 2020; Trebing et al., 2021) also confirmed that the U-Net architecture can accurately predict precipitation.

In the field of computer science, the generative adversarial network (GAN) architecture (Goodfellow et al., 2014) showed remarkable performance in image-to-image tasks. Isola et al. (2017) demonstrated that the U-Net model with a conditional GAN (cGAN) approach called Pix2Pix can generate higher quality images than the original U-Net model. Rüttgers et al. (2019) showed that typhoon tracks and cloud patterns over the Korean Peninsula could be successfully predicted using cGAN architecture with satellite cloud images. Also, Ravuri et al. (2021) developed a precipitation nowcasting model using a deep generative model inspired by the video GAN model (Clark et al., 2019). In the case study of convective cells over eastern Scotland, using video GAN in the model improved the quality of precipitation forecasts significantly (Ravuri et al., 2021). These studies indicate that the performance of precipitation nowcasting models can be improved by advanced machine-learning techniques. However, because machine-learning is a data-driven technique, it will perform effectively only for trained data domains. Generally, it is vital to train from the beginning to develop a model for a new domain, and computation costs will be high even if new data are similar to old data. Thus, the models trained for one domain will be limited in their applications for multiple regions.

The aim of this study was to develop an advanced precipitation nowcasting model for multiple dam basins that can be applied as an early warning system. The decision-making process at upstream dams with regard to flood control, which is directly related to urban and rural water management, influences flood risk considerably. From such a dam management perspective, water level and inflow at dam sites are major factors to be considered, suggesting that increasing rainfall prediction accuracy over the entire dam basin is essential for effective flood management. To develop an advanced precipitation nowcasting model with good prediction performance for dam basins in general, we designed a model based on the cGAN approach (Rad-cGAN) for multiple dam domains of the Soyang-gang, Andong, and Chungju dam basins in South Korea. We trained the model using radar reflectivity data from the Soyang-gang Dam basin for the summer season during 2014–2017 (provided by the Korea Meteorological Administration, KMA) and evaluated model performance using the 2018 data by comparing it with reference models of ConvLSTM, U-Net, and Eulerian persistence. We also used spectral prognosis (S-PROG) (Seed, 2003), which is a deterministic nowcast model in the pySTEPS library, for evaluation. Then, we applied the transfer learning technique (Pan and Yang, 2010), which uses the previously trained model with cost-effective computation to train the model for the other two abovementioned domains. Five transfer learning strategies were compared to evaluate the most effective cases for model development for the Andong Dam and Chungju Dam basins.

2 Materials and methods

2.1 Study area and radar reflectivity data

We developed a precipitation nowcasting model for dam basins where an accurate rainfall forecasting system is essential for the estimation of urban water supply and flood prevention. The target domains are the Soyang-gang Dam basin (D1), Chungju Dam basin (D2), and Andong Dam basin (D3) areas. These dams are multi-purpose and are located upstream of the major rivers of South Korea (Fig. 1).

https://gmd.copernicus.org/articles/15/5967/2022/gmd-15-5967-2022-f01

Figure 1(a) Composite map of radar reflectivity and location of the dam basins; (b) selected areas over the dam basin. D1, D2, and D3 represent the areas of the Soyang-gang, Chungju, and Andong dam basins respectively. Maps were created using ArcGIS software by Esri; Base-map source: Esri, HERE, Garmin, © OpenStreetMap contributors, and the GIS User Community.

The 1.5 km constant altitude plan position indicator (CAPPI) radar reflectivity data, provided by KMA, were used as input data for the training and evaluation of our model. The map product represents the quality-controlled radar reflectivity composite (dBZ) of 11 weather radar stations across South Korea (Fig. 1a), with a size of 960 × 1200 pixels, spatial resolution of 1 km, and temporal resolution of 10 min.

The radar composite data were cropped to 128 × 128 pixels, covering three target basins. Figure 1b shows the different topographical characteristics of each domain. As topography (especially mountainous areas such as study domains) affects atmospheric conditions, such as temperature, humidity, air pressure distribution, and cloud formation, it directly or indirectly affects rainfall formation and distribution (Basist et al., 1994; Prudhomme and Reed, 1998). Consequently, data extracted from the three domains with different topographic characteristics would exhibit different rainfall patterns. We selected the available radar reflectivity data in summer (June–August, JJA) from 2014 to 2018 considering that high-intensity rainfall occurs in summer owing to rainfall seasonality, a characteristic of our study domain. Data from 2014 to 2017 were used to train the model, and data from 2018 were used for evaluation (Table 1). For rapid and effective training, the raw radar reflectivity data (dBZ) were converted to grayscale (0–255), and the data range was scaled to 0–1 using the Min–Max scaler method (min–max values from the training dataset). The predicted radar reflectivity data were converted into precipitation using the ZR relationship (Marshall and Palmer, 1948) to evaluate the rainfall prediction performance of the model:

(1) Z = 200 R 1.6 ,

where Z is the radar reflectivity factor (mm6 m−3) and R is the rainfall rate (mm h−1).

Table 1Distribution of precipitation amount and the number of examples of (a) the training dataset, and (b) the test dataset of each dam basin.

Download Print Version | Download XLSX

We used cropped radar reflectivity images (128 × 128 km2) of the Soyang-gang Dam basin to train and evaluate the proposed model (Rad-cGAN) and reference models (U-net, ConvLSTM, Eulerian persistence, and pySTEPS (S-PROG)). Furthermore, to reduce the edge effect caused by the fast Fourier transform (FFT), which is used for scale decomposition of the pySTEPS (S-PROG) nowcast (Pulkkinen et al., 2019; Foresti and Seed, 2014), we derived the pySTEPS results using 384 × 384 km2 input data extended by 128 pixels on each side of the original input data (128 × 128 pixels).

2.2 Model architecture

2.2.1 Conditional generative adversarial network for image translation

Generative adversarial network (GAN) is a recently developed framework for training generators (e.g., CNN encoder-decoder) via an adversarial process. It consists of a generator (G) that produces the distribution of real data from random noise, and a discriminator (D) that classifies whether the input sample is from the generator or the original data distribution (Goodfellow et al., 2014). Furthermore, the conditional generative adversarial network (cGAN) framework uses additional conditions (e.g., input data of the generator) for training and can generate targeted outputs that suit specific conditions (Mirza and Osindero, 2014). For image translation tasks, when G is trained to produce a targeted image (y) from input (x) with random noise (z), the objective of D will try to maximize the loss function LcGAN(G,D) whereas G will try to minimize LcGAN(G,D). This relation can be expressed as:

(2) min G max D L cGAN G , D = E x , y log D x , y + E x , z log ( 1 - D x , G ( x , z ) ,

where losses were calculated as expected (E) values. After simultaneously training G and D, G was trained to generate an output that cannot be distinguished from real data (y) by D, which was trained in an adversarial manner to detect the fake image from G. Isola et al. (2017) showed that combining the traditional pixel-wise loss with cGAN loss can improve the quality of output images. To generate sharp and realistic images, the L1 loss function LL1(G) was used as the traditional loss:

(3) L L 1 G = E x , y [ y - G x , z 1 ] .

By adding the traditional loss with a weight λ to the cGAN loss, the final objective was obtained:

(4) G = arg min G max D L cGAN G , D + λ L L 1 ( G ) .

In this study, we developed a radar-based precipitation nowcasting model using a cGAN framework. Recently, research on weather prediction using cGAN, an advanced machine-learning approach, has been conducted extensively (e.g., Rüttgers et al., 2019; Ravuri et al., 2021). For example, Ravuri et al. (2021) proposed a generator consisting of two modules; conditioning stack (using CNN to extract a representation of input); and sampler (using ConvGRU to generate prediction). The model, which used ConvGRU, could observe spatiotemporal changes of inputs, such as ConvLSTM, and attempted to improve performance by extracting features from different spatial dimensions and deriving the results. Whereas the generator was used to predict future radar maps, the discriminator used a dual architecture that distinguishes the real and generated frames, to ensure both temporal and spatial consistency. Unlike the model proposed by Ravuri et al. (2021), our model adopts a U-net architecture that uses a CNN layer in image generation based on the underlying Pix2Pix model; the architecture exhibits outstanding performance in image-to-image translation tasks (Isola et al., 2017). Also, we considered only spatial consistency in the PatchGAN discriminator, which distinguishes images for each N×N patch (N can be smaller than the full size of the image). The U-net-based precipitation nowcasting model has previously demonstrated performance superior to that of a traditional radar-based precipitation nowcasting model that uses optical flow (Ayzel et al., 2020). Therefore, here, we apply the basic cGAN methodology to the U-net structure to improve performance and confirm the applicability of the transfer learning methodology to multiple dam domains.

2.2.2 Generator

Figure 2a shows the generator using U-Net architecture (a detailed description of U-Net is provided in Sect. 2.3.3). The model consists of nine convolutional layers, two max-pooling layers, two up-sampling layers, and an output convolutional layer. Each convolutional layer, except for the output layer, is composed of the following operations: 3×3 2D convolution with zero padding, batch normalization, and activation function of ReLU. In the contracting part of the generator, a 2×2 2D max-pooling operation was used to down-sample the input images. To prevent overfitting, a dropout layer with a rate of 0.5 was applied after the pooling and convolutional layers of the expanding part of the model (Srivastava et al., 2014). A 2×2 2D up-sampling operation was further applied in the expanding part after skip connection to increase the resolution of featured images that contain both high- and low-level information. Finally, the output convolutional layer had a 1×1 2D convolution that used a linear function for activation to obtain a future prediction of the radar reflectivity image.

2.2.3 Discriminator

PatchGAN from the Pix2Pix model was used as the discriminator (Fig. 2b). As in cGAN, the input pair of the discriminator consists of historical radar reflectivity data (i.e., input of the generator) and future radar reflectivity data. The discriminator classifies real image pairs (input of the generator and ground truth image) as 1 and fake image pairs (input and generated image from the generator) as 0 (Mirza and Osindero, 2014). In particular, PatchGAN only penalizes the structures over a certain scale of image patches; therefore, the discriminator classifies whether the N×N patch in the input pair is real or fake. This patch represents the receptive field, which is the region in the input image that is used to measure the associated feature of the output layer. Consequently, the size of the patch (N) was determined based on the structure of the entire discriminator (e.g., number of layers, nodes, filter size, paddings, and strides), and it increased as the model deepened. We constructed a discriminator model with a 34×34 patch size through hyperparameter tuning. The model consists of three convolutional layers and an output layer. The first two convolutional layers were composed of 4×4 2D convolution with strides of two and zero padding, batch normalization, and ReLU activation function, which was leaky and had a 0.2 slope. The third convolutional layer had the same configuration as the previous layers, except that its stride was 1. To distinguish the input pair in the image form, the output layer consisted of 4×4 2D convolution with zero padding and a sigmoid activation function. To train the discriminator as a classifier, we manually generated the training datasets consisting of the input image pairs and the target images, with spatial dimensions of 32×32 filled with 1 (for real image pairs) or 0 (for generated image pairs). Therefore, each pixel of the output estimates the probability that the discriminator determines each patch of the input pair as the real one.

https://gmd.copernicus.org/articles/15/5967/2022/gmd-15-5967-2022-f02

Figure 2Model architecture consists of (a) the generator and (b) the discriminator.

Download

2.2.4 Optimization procedure

Before proceeding with training to optimize the model for the input data, hyperparameter tuning is required to determine the optimal model structure and training settings. We selected the following hyperparameters: number of layers, number of hidden nodes, convolution filter size, patch size, batch size, and learning rate. To select the appropriate hyperparameter combination, the model for each combination was trained using radar data from 2014 to 2016 (June to August) and data from 2017 (June to July). Subsequently, using data from 2017 (August), the mean absolute error (MAE) and critical success index (CSI) (at an intensity threshold of 0.1 mm h−1) were calculated to obtain the optimal combination of hyperparameters. Based on the tuning results, the MAE range was 0.45–47.66 and the CSI range was 0.0–0.83, and the results confirmed that hyperparameters influence model performance considerably. Based on the combinations that performed optimally, we determined the model structure and training settings.

To optimize Rad-cGAN, the training procedure suggested by Isola et al. (2017) was adopted. First, we compared the results using a total of four and six consecutive radar reflectivity images to determine the input historical data length. As a result of a 10 min precipitation prediction at the Soyang-gang Dam site, in the case of CSI (at the rainfall intensity of 0.1 mm h−1), the case of using six historical data was slightly better than the case of using four historical data, but in R, RMSE, and NSE, the results of using four data were better. Through this, samples that consisted of four consecutive radar reflectivity images (t−30, t−20, t−10 min, and t) and the image at t+10 min were selected to train the model. Subsequently, the training samples for the discriminator were created by adding labels to classify whether the samples were real (image at t+10 min from observation) or fake (t+10 image from the generator) pairs. Then, the parameters of the discriminator were updated using the minibatch stochastic gradient descent (SGD) method for one step. Binary cross-entropy was used as a loss function, and the ADAM optimizer (Kingma and Ba, 2015) with a learning rate of 0.0002 and momentum parameters β1=0.5 and β2=0.999 was applied. Afterwards, the generator was trained for one step to optimize Eq. (4). Binary cross-entropy was used as LcGAN for the discriminator to classify the generated image into a real image. Additionally, λ of the traditional pixel-wise L1 loss was set to 100. The minibatch SGD and ADAM optimizer were applied to train the generator with the same setting as the discriminator. Both the procedures for updating the parameters of the discriminator and generator were run simultaneously during one epoch. Our model was trained using 600 epochs, with a batch size of 8. To achieve the optimal model, we applied an early stopping technique that stops the training model when the loss stops improving. The loss metric was defined as the generator loss based on 100 validation samples randomly sampled from the training dataset that was not used to train the model. To monitor the loss, we set patience to 30 epochs and saved the model when the loss improved. The model architecture was written in Python (https://www.python.org/, last accessed: 30 June 2022) using the Keras deep learning application (https://keras.io/, last access: 30 June 2022). The entire procedure for training and evaluating experiments was run on a computer with a single NVIDIA Tesla V100 GPU.

2.3 Reference models

The performance of Rad-cGAN was compared with and validated using reference models that include two baseline and two machine-learning models. We used the Eulerian persistence model (hereafter referred to as Persistence), a traditional radar-based rainfall prediction model, as the baseline model. This model assumes that rainfall prediction at any lead time is the same as the rainfall in the forecast time. This is a simple but powerful model for predicting short-term precipitation. Additionally, a deterministic S-PROG nowcast from the pySTEPS library was used as another baseline model (Sect. 2.3.1). We used ConvLSTM (Sect. 2.3.2) and U-Net (Sect. 2.3.3), which are the common basic structures for machine-learning-based nowcasting models, as reference models for comparison.

2.3.1 PySTEPS

PySTEPS (Pulkkinen et al., 2019) is an open-source and community-driven Python framework for radar-based deterministic and probabilistic precipitation nowcasting and is considered a strong baseline model (Imhoff et al., 2020; Ravuri et al., 2021). In this study, deterministic S-PROG (Seed, 2003) nowcast from the pySTEPS library was used as the benchmark model.

To predict precipitation, we input the precipitation images (unit: dBR) transformed from four consecutive radar reflectivity images (from t−30 to t), which were the same as the input of Rad-cGAN, based on the ZR relationship (Eq. 1). Additionally, the transformed precipitation was used to estimate the motion field, which was used together with precipitation as input data in the model. Future precipitation at a lead time of up to 90 min for the test period (JJA, 2018) was generated from the results of the S-PROG nowcasts. The source code of pySTEPS is available at GitHub repository (https://pysteps.github.io, last access: 23 May 2022).

2.3.2 ConvLSTM

LSTM is a special case of recurrent neural networks (RNNs) and is widely used in temporal sequence predictions (Hochreiter and Schmidhuber, 1997). Sutskever et al. (2014) proposed an LSTM encoder-decoder framework for sequence-to-sequence problems, which consists of concatenated LSTMs for the input and output sequences. Based on this model, Shi et al. (2015) developed a ConvLSTM network that can be applied to spatiotemporal sequence prediction, such as radar-based rainfall prediction. To handle spatiotemporal sequences, a convolution operator was used in state-to-state and input-to-state transitions. The ConvLSTM model was shown to outperform the traditional optical flow-based precipitation nowcasting model. Recent studies have shown that the ConvLSTM model can be successfully applied to predict future radar-based precipitation (Kim et al., 2017; Moishin et al., 2021).

We designed a ConvLSTM model that uses four radar reflectivity image frames (t−30, t−20, t−10 min, and t) as input to predict future frames at time t+10 min, which is similar to the input and output of Rad-cGAN. The model consists of three ConvLSTM layers and an output layer. Each ConvLSTM layer contains 64 hidden states and 3×3 kernels. A 3D convolutional layer with a linear activation function was used as the output layer. The hyperparameters of the ConvLSTM model (i.e., number of layers, number of nodes, convolution filter size, batch size, and learning rate) were tuned using a procedure similar to that applied in Rad-cGAN (Sect. 2.2.4). To optimize the model, we used the mean squared error as the loss function and applied the ADAM optimizer (learning rate 0.002 and momentum parameters β1=0.9 and β2=0.999). We trained the model using 600 epochs (early stopping applied) with a batch size of 32.

2.3.3 U-net

U-Net-based precipitation nowcasting models efficiently predict future precipitation using historical data, even though U-Net does not have a structure, such as RNN, that preserves temporal information (e.g., Ayzel et al., 2020; Trebing et al., 2021). U-Net was developed by modifying the fully convolutional network (Long et al., 2015), and performed well in image segmentation tasks (Ronneberger et al., 2015). This model architecture consists of two parts: a contracting network that captures the context of the input images and an expanding network that increases the resolution of features from the contracting network.

The contracting network follows the usual CNN, which consists of convolutional layers and max-pooling layers. Each convolutional layer is composed of convolution, batch normalization, and activation operations. Batch normalization is used to prevent gradient vanishing or exploding problems and can effectively increase the convergence speed (Ioffe and Szegedy, 2015). The max-pooling operation is applied for down-sampling after the convolution of the input image. Through this process, the output of the contracting network can incorporate the features of the input image. The expanding network consists of the up-sampling and convolutional layers. Before applying the up-sampling operation, the skip connection is applied between each layer of the contracting network and the layer of the expanding network to prevent gradient vanishing and share the low-level information of the input data (Simonyan and Zisserman, 2015). The convolution layers of the expanding and contracting networks follow the same operation.

As the reference model, hyperparameters for the U-net structure (number of layers, number of nodes, and convolution filter size) were set to be equivalent to those of Rad-cGAN (Sect. 2.2.2), and hyperparameters related to training settings (batch size and learning rate) were tuned using procedures similar to those of Rad-cGAN (Sect. 2.2.4). To optimize the model, L1 loss and ADAM optimizers were used as in the case of ConvLSTM (Sect. 2.3.2). The model was trained using 600 epochs with early stopping and the batch size set to 8.

2.4 Experiments for evaluating the model's prediction skills

2.4.1 Performance evaluation

The model was trained using data from the summers (June–August) of 2014–2017 and its precipitation nowcasting capacity was assessed using data from the summer of 2018. To predict radar reflectivity data 10 min ahead, four latest radar reflectivity data (t−30, t−20, t−10 min, and t min; t being the forecast time) were used as input data. The model can generate multiple samples (number of samples, 128, 128, 1) corresponding to the number of samples of the past four consecutive input data (number of samples, 128, 128, 4). To predict beyond the 10 min lead time, we used the prediction data at t+10 min as the latest input data. Using this recursive process, predictions were obtained at a lead time of >10 min. Because the model predicts the radar reflectivity after 10 min using past consecutive radar images, we first evaluated the model performance at a lead time of 10 min. This allowed us to confirm the prediction tendency of our model and other reference models while performing precipitation nowcasting. Furthermore, to assess the applicability of our model to the actual early warning system that needs to ensure at least 1 h of lead time, the predictive skill was evaluated for >10 min of lead time using the recursive process. We measured the verification metrics (see below) using rainfall prediction, converted from the radar reflectivity (Eq. 1), at a lead time of up to 90 min to confirm the forecasting time in which the model ensured sufficient performance.

We evaluated the model performance of the entire dam basin as the water level at the dam site by rainfall over the dam basin is a major factor in the decision-making process for dam management (Fig. 1b). To evaluate the entire domain, the verification metrics were calculated with increasing lead time for all pixels of the predicted image. Additionally, to qualitatively evaluate the model performance, we compared the resulting precipitation images obtained using data at a certain forecast time. We set the forecast time at 23 August 2018, 17:50 UTC, when typhoon Soulik, which landed on the Korean Peninsula from 23 August 2018, 12:00 UTC to 24 August 2018, 03:00 UTC, started affecting the Soyang-gang Dam basin.

Several metrics were used for model evaluation: Pearson correlation coefficient (R), root mean square error (RMSE), Nash–Sutcliffe efficiency (NSE), CSI, and fractions skill scores (FSS). As the collinearity between actual rainfall and predicted rainfall increases, the explanatory power of the rainfall simulated by the model increases, so that the performance of the model can be illustrated by the strong positive linear relationship between predictions and observations. Hence, we confirmed that the model exhibits better performance when R (Eq. 5), calculated based on the model prediction and observation, is closer to 1. To verify the precision of the model, the RMSE (Eq. 6) between prediction and observation was used. In addition, NSE, widely used to assess hydrologic models, was used as a goodness-of-fit index for the proposed precipitation nowcasting model (McCuen et al., 2006) (Eq. 7).

(5)R=i=1N(Oi-O)(Pi-P)i=1N(Oi-O)2i=1N(Pi-P)2,(6)RMSE=i=1N(Oi-Pi)2N,(7)NSE=1-i=1N(Oi-Pi)2i=1N(Oi-O)2,

where O and P are the means of observation and prediction respectively, Oi and Pi are the observed and predicted precipitation respectively, in the ith time of the data period, and N is the total number of data for the entire period.

We used the CSI (Eq. 8), which is a measure of categorical forecast performance, to verify the model accuracy for precipitation event detection:

(8) CSI = hits hits + false alarms + misses ,

where hits (correct event forecasts), false alarms (incorrect event forecasts), and misses (missed events) are defined by a contingency table (Table 2). Also, FSS can spatially verify model performance by comparing the fraction of grid points of prediction and ground truth, which exceed certain rainfall intensity thresholds within the neighborhood:

(9) FSS = 1 - i = 1 n ( P p - P o ) 2 i = 1 n P p 2 + i = 1 n ( P o ) 2 ,

where Pp and Po are the fractions of prediction and observation respectively, calculated by specific thresholds in neighborhood size. For calculating CSI and FSS, we selected several intensity thresholds, including 0.1, 1.0, and 5.0 mm h−1, and for FSS, we used neighborhood sizes of 1, 5, and 15 km. Additionally, we calculated the radially averaged power spectral density (PSD) of predictions and observations to assess the blurring effect of the predicted images by models.

To calculate each verification metric, all metrics for each pixel in the dam basins were calculated and averaged over the data period (number of samples).

Table 2Contingency table for the categorical scores.

Download Print Version | Download XLSX

2.4.2 Experiments for transfer learning among different domains

As the machine-learning model relies on input data as a data-driven model, training on the corresponding new data must be conducted from the beginning to develop a model for a new domain, which is also applicable for our precipitation nowcasting model for a new dam basin with different meteorological, environmental, and geographical characteristics (Fig. 1b). However, because this method is time-consuming and computationally expensive, we applied a transfer learning approach that can be efficiently used to train models with multiple dam basins.

Transfer learning is a machine-learning technique that uses knowledge and skills from the pre-trained models to train a model for new datasets (Pan and Yang, 2010). This method is often used when the size of the provided dataset is insufficient for training and is also used to train the models for the new dataset owing to its lower computational cost than that of training from scratch.

The general training strategies of transfer learning are determined by the data size and similarity between the new and the original data. For example, if the new dataset is similar to the dataset of the pre-trained model, the new model only fine-tunes for higher layers that learn specific features of the input data and freeze the lower layers that capture the general features. Fine-tuning uses a smaller learning rate (e.g., 1/10th of the original learning rate) and is one of the most effective ways of transferring knowledge. Several studies have shown that the transfer learning approach performs successfully well in image classification tasks (Krizhevsky et al., 2012; Simonyan and Zisserman, 2015; He et al., 2016). In the GAN approach, the discriminator acts similar to the classifier of the image classification task. Wang et al. (2018) reported that fine-tuning both the generator and discriminator resulted in a good performance, but overfitting was a frequent issue that must be considered. Subsequently, Mo et al. (2020) proposed a strategy that works only on the discriminator called FreezeD, which freezes the lower layer of the discriminator and only fine-tunes the upper layers.

We used transfer learning to train our model for different dam basins, i.e., Andong and Chungju, with a pre-trained model that was completely trained by data from Soyang-gang Dam basin. In addition, existing papers that successfully applied the transfer learning strategies developed the models for the new domains using the pre-trained model based on vast data. Consequently, we used the pre-trained model with Daecheong Dam, Juam Dam, and Yongdam Dam basin data, in addition to Soyang-gang Dam data, to assess the amount of data required to develop a model for a new dam domain. The selected strategies were inspired by a previous approach to transferring GAN (Wang et al., 2018; Mo et al., 2020). We formulated two strategies for each pre-trained model. First, the weights of the pre-trained generator were frozen and used directly in the new dam domain (Cases 2 and 4). Next, the weights of the pre-trained generator were fine-tuned (1/10th of the original learning rate) and the discriminator was trained from scratch (Cases 3 and 5). In addition, the entire model was trained for the new domain (Case 1) (Table 3a). The model was trained for the Chungju Dam and Andong Dam domains, separately, using the five strategies (Table 3b). To determine the best strategy for training different dam domains, we estimated the performance at the 10 min lead time at each dam domain (Fig. 1b). Additionally, we compared the predictive skill of each strategy at the lead time of up to 90 min by using the recursive process.

Table 3Experimental design for transfer learning strategies to train the model with different domains. (a) Detailed training procedure of each strategy, (b) data used to train the model according to each strategy.

* Use 1/10th of original learning rate.

Download Print Version | Download XLSX

3 Results and discussion

3.1 Domain-averaged model performance for the Soyang-gang Dam basin

To apply our model in an early warning system, the rainfall prediction performance upstream of the dam should be sufficient. Hence, the verification metrics were calculated for each grid cell in the entire domain. First, we evaluated the performance of our model for the predicted precipitation at a lead time of 10 min at the Soyang-gang Dam basin during the summer of 2018 (Table 4). As the general criterion for evaluating hydrological models, when R and NSE are  0.5, the model has acceptable performance (Moriasi et al., 2007). In addition, Germann and Zawadzki (2002) suggested that the threshold of predictability is 1/e (about 0.37), assuming that the verification metrics follow the exponential law. According to the standard, the mean values of each metric in Table 4 show that the machine-learning-based models generally performed well as precipitation nowcasting models (R>0.5, NSE > 0.5, CSI > 0.5). Among them, Rad-cGAN outperforms the other reference models for almost all the verification metrics (Table 4). Particularly, Rad-cGAN shows improvements in the CSI values at different rainfall intensities (0.1, 1.0, and 5.0 mm h−1) by 0.55 %, 10.10 %, and 123.50 % respectively, compared with the model results using U-Net, confirming that the cGAN approach can mitigate the tendency to underestimate precipitation. However, by comparing with ConvLSTM and baseline models (especially for pySTEPS), Rad-cGAN performs poorly for the highest rainfall intensity (5.0 mm h−1).

Table 4Comparison of the average values of the verification metrics for the 10 min precipitation prediction of different models at the Soyang-gang Dam basin during summer (June–August) 2018.

Download Print Version | Download XLSX

We predicted the precipitation at the lead times of up to 90 min by using the recursive process, and their performances in all the grid cells were presented through boxplots for each lead time in all the models (Fig. 3). By comparing the median values, Fig. 3 shows average increases of 9.02 % and 17.87 % for the R of Rad-cGAN at overall lead times compared with those of U-Net and ConvLSTM respectively, which indicate improved precipitation prediction capacity for the entire domain. However, in the cases of RMSE and NSE, Rad-cGAN performs slightly worse than ConvLSTM, with an average increase over median values of 1.90 % in RMSE and a decrease of 7.67 % in NSE over the entire lead time (Fig. 3).

https://gmd.copernicus.org/articles/15/5967/2022/gmd-15-5967-2022-f03

Figure 3Box plots of the verification metrics of model predictions at the lead time up to 90 min over all grid cells from the Soyang-gang Dam basin. From top to bottom, (a)(c) represent R, RMSE, and NSE, and (d)(f) represent the CSI at intensity thresholds of 0.1, 1.0, and 5.0 mm h−1. Dotted gray line represents the predictability threshold (1/e).

Download

Moreover, according to the CSI value at the intensity of 0.1 mm h−1, our model preserves its predictability performance (>1/e) for the entire lead time, indicating that it can be applied to predict precipitation at lead times of >90 min. By comparing with previous studies, the lead time for the CSI at the intensity of 0.1 mm h−1> 0.5 was up to 90 min with ConvLSTM in this study, whereas CSI at the intensity of 0.5 (not 0.1) mm h−1> 0.5 was up to 40 min with ConvLSTM-based nowcasting model for Hong Kong region (Shi et al., 2015). Also, Ayzel et al. (2020) showed that the U-net-based model preserved performance (CSI at an intensity of 0.125 mm h−1> 0.5) at a lead time of >60 min in Germany, whereas the performance of our model with similar CSI (0.1 mm h−1> 0.5) remained up to 90 min. Hence, we confirm that the reference models were sufficiently trained to be used for comparison with our model.

The results of CSI indicate that Rad-cGAN has a reliable performance in precipitation nowcasting for relatively light rain (rainfall intensity of 0.1 mm h−1). However, in the case of CSI at intensities of 1.0 and 5.0 mm h−1, although Rad-cGAN maintains a good performance compared with that of U-net, the performance rapidly deteriorates as the lead time increases. Unlike Rad-cGAN and U-net, ConvLSTM and baseline models record low CSI under low-intensity rainfall; however, Fig. 3 shows that relatively high levels of performance are maintained under higher rainfall intensity. These results can also be confirmed through the FSS of each model (Fig. 4). Rad-cGAN is better when comparing Rad-cGAN and U-net, but both models significantly decrease FSS as lead time and rainfall intensity increase. However, ConvLSTM and pySTEPS have relatively high FSS values under high rainfall intensity compared with those of the other two models.

https://gmd.copernicus.org/articles/15/5967/2022/gmd-15-5967-2022-f04

Figure 4Fraction skill scores (FSS) of model predictions at lead times of 10, 30, and 60 min at Soyang-gang Dam basin. Panels from left to right express FSS of Rad-cGAN, U-net, ConvLSTM, and pySTEPS (S-PROG).

Download

Thus, we observe a tendency to underestimate the prediction of high-intensity precipitation in all models, including Rad-cGAN (Figs. 3 and 4). A similar observation was made using ConvLSTM (Kumar et al., 2020), wherein significant errors occurred in precipitation prediction (>20 mm h−1). This may be attributed to data imbalance, which is a common issue in machine-learning studies (Wang et al., 2016). Data imbalance occurred in this study because, unlike low-intensity precipitation (<5 mm h−1), high-intensity precipitation rarely occurs during the training and testing periods.

3.2 Spatial model performance for the Soyang-gang Dam basin

To better understand model performance with increasing lead time, we predicted precipitation for lead times of 10, 30, 60, and 90 min for a specific forecast time on 23 August 2018, 17:50 UTC, when typhoon Soulik began affecting the Soyang-gang Dam basin (Fig. 5). We observe that with an increase in lead time, the model performance deteriorates owing to the blurring effect of the predicted image, which is an issue reported in previous machine-learning-based nowcasting models (Ayzel et al., 2020; Shi et al., 2015). Despite the smoothing trend, Rad-cGAN produces qualitatively better results than those of the other reference models (Fig. 5). The bias of prediction (observation-prediction) of Rad-cGAN at the 90 min lead time ranges from −1.97 to 19.68 (mean = 0.83 mm h−1), indicating that our model alleviates the underestimation of precipitation compared with U-Net, whose bias ranges from −0.30 to 20.33 (mean = 1.04 mm h−1). These results support the improvement in Rad-cGAN verification metrics compared with those of U-Net (Figs. 3 and 4). Furthermore, in the case of ConvLSTM, the mean bias of 0.86 mm h−1 under the 90 min lead time prediction shows that ConvLSTM is less prone to underestimation compared with U-Net. However, the 90 min rainfall prediction by ConvLSTM is recorded to be approximately 0 mm h−1 in areas with an observation of  5–10 mm h−1, indicating that it predicts precipitation to be close to zero in most areas with increasing lead times. As Fig. 5 illustrates, ConvLSTM does not predict the boundaries of the entire precipitation area well but maintains a higher intensity rainfall than the other two machine-learning-based models, which causes the CSI difference according to the rainfall intensity of ConvLSTM (Figs. 3 and 4).

https://gmd.copernicus.org/articles/15/5967/2022/gmd-15-5967-2022-f05

Figure 5Example of precipitation at forecasting time t= 23 August 2018, 17:50 UTC, for model predictions and ground truth (observation). Panels from top to bottom express (a) ground truth, (b) prediction of Rad-cGAN, (c) prediction of U-net, (d) prediction of ConvLSTM, and (e) prediction of pySTEPS.

Figure 6 shows the PSD for each result in Fig. 5. Based on Figs. 5 and 6, all models exhibit the blurring effect compared with the ground truth. However, when comparing U-net and Rad-cGAN, Rad-cGAN has a slightly lower blurring effect (Fig. 6). This is because sharper images can be generated when cGAN is applied to the U-net structure (Isola et al., 2017), indicating that the cGAN technique was successfully applied by our model. Therefore, based on the overall verification metrics, we conclude that Rad-cGAN has the optimal performance in nowcasting and prediction of spatial patterns of movement of precipitation.

https://gmd.copernicus.org/articles/15/5967/2022/gmd-15-5967-2022-f06

Figure 6Radially averaged power spectral density at forecasting time t= 23 August 2018, 17:50 UTC, for model predictions and observations.

Download

PySTEPS shows poor performance compared with previous studies (e.g., Imhoff et al., 2020) in the verification metrics (Table 4 and Figs. 3–5). The overall prediction performance degrades particularly because the precipitation area near the edge of the basin is not predicted (Fig. 5). To better understand this edge effect, we reran pySTEPS and Rad-cGAN with the extended data of 384×384 pixels. Compared with the predictions in Fig. 5, the typhoon event prediction from each model (Fig. 7) shows that using the extended area reduces the edge effect of pySTEPS and properly maintains high rainfall intensity, thereby improving the performance. Moreover, the average R and CSI (at the highest rainfall intensity of 5.0 mm h−1) for the 10 min precipitation prediction during the entire test period are calculated as 0.77 and 0.38 respectively, indicating that the performance improves quantitatively compared with the previous results (R=0.70 and CSI = 0.32). Additionally, the prediction performance of typhoon events improves in Rad-cGAN using the extended area (Fig. 7), and the average R and CSI (at the rainfall intensity of 5.0 mm h−1) in the 10 min rainfall prediction for the entire test period improve from 0.79 to 0.80 and from 0.18 to 0.37 respectively. Both models show improved performance using extended area, but considering the applicability of the model to real-world problems with limited data availability, we conclude that Rad-cGAN, unlike pySTEPS, is more efficient at predicting rainfall without considering the edge effects due to the spatial size of the input domains.

https://gmd.copernicus.org/articles/15/5967/2022/gmd-15-5967-2022-f07

Figure 7Example of precipitation at forecasting time t= 23 August 2018, 17:50 UTC, for model predictions using increased input area (384 × 384). Panels from top to bottom express (a) prediction of Rad-cGAN, and (b) prediction of pySTEPS.

3.3 Performance with transfer learning at different dam domains

To develop the precipitation nowcasting models for multiple dam basins (Andong and Chungju dam basins) other than Soyang-gang Dam basin, we proposed to retrain our model with data from new dam basins (Case 1) as well as apply the efficient transfer learning methodologies (Cases 2–5; Sect. 2.4.2).

First, we evaluated whether the transfer learning approach could be effectively applied to the new domain (Cases 2 and 3) using a pre-trained model only for one domain (i.e., Soyang-gang Dam basin). Table 5a shows the performance of each case with model-predicted precipitation at a 10 min lead time at the Andong Dam basin. The results show that most of the verification metrics in Case 2 and Case 3 perform better than those in Case 1. In Case 2, which uses all the parameters of the generator from the pre-trained model, an NSE of 0.56 is achieved, which is closest to the NSE of the pre-trained model (0.54) with data from the Soyang-gang Dam basin, indicating that Case 2 achieves sufficient performance for the new domain. These results are consistent with the verification metrics at lead times of up to 90 min (Fig. 8). Based on the median of R, Case 2 maintains the predictive performance (>1/e) up to about 80 min lead time (Fig. 8a). Especially for CSI at higher rainfall intensities, Case 2 shows better performance than Case 1 overall lead times (Fig. 8d–f). Hyperparameter tuning would have had a significant impact on the results where Case 2 performs better than Case 1. Unlike the pre-trained model, which confirmed that model optimization and generalization were completed through the hyperparameter tuning process, in Case 1, we did not proceed with hyperparameter tuning for the new domain. Although the new domain has properties similar to those of the pre-trained domain, minor changes in hyperparameters also result in differences in performance, so that optimization and generalization of the model (Case 1) were less comprehensive than in the pre-trained model, resulting in relatively poor performance. However, in Case 3, which used the same pre-trained model as Case 2, the performance is lower than that of the other strategies. This is because of performance degradation due to overfitting during fine-tuning the pre-trained parameters. Great similarity between the two datasets of the Andong and Soyang-gang Dam domains may be the reason for major performance degradation of the transfer learning using the fine-tuning method (Wang et al., 2018).

https://gmd.copernicus.org/articles/15/5967/2022/gmd-15-5967-2022-f08

Figure 8Box plots of the verification metrics of model predictions at the lead time up to 90 min over all grid cells from the Andong Dam basin. Panels (a)(c) represent R, RMSE, and NSE, and (d)(f) represent CSI at intensity thresholds of 0.1, 1.0, and 5.0 mm h−1 respectively. Dotted gray line represents the predictability threshold (1/e).

Download

Table 5Comparison of the average values of the verification metrics for the 10 min precipitation prediction of the five different models using different transfer learning strategies for the (a) Andong Dam and (b) Chungju Dam basins in summer (June–August) 2018.

Download Print Version | Download XLSX

For the Chungju Dam basin, we trained the model using the same methodologies as those used for the Andong Dam basin. Among Cases 1, 2, and 3, transfer learning cases (Cases 2 and 3) perform better than Case 1 for the overall verification metrics, especially for CSI at higher rainfall intensities (Table 5b). Additionally, when the lead time is increased by up to 90 min, Cases 2 and 3 show better performance than Case 1 for the entire lead time (Fig. 9). From the median values of CSI at the 0.1 mm h−1, all three cases preserve sufficient performance (CSI >1/e) at a lead time of up to 90 min. However, as a result of comparing CSI with higher rainfall intensity (1.0 mm h−1), it is confirmed that the performance in Case 2 only maintains predictive performance for up to 30 min (Fig. 9e). Through these results, Case 2 is successfully applied to Andong Dam and Chungju Dam basins among the transfer learning strategies using the pre-trained model for the Soyang-gang Dam basin.

https://gmd.copernicus.org/articles/15/5967/2022/gmd-15-5967-2022-f09

Figure 9Box plots of the verification metrics of model predictions at the lead time up to 90 min over all grid cells from the Chungju Dam basin. Panels (a)(c) represent R, RMSE, and NSE, and (d)(f) represent CSI at intensity thresholds of 0.1, 1.0, and 5.0 mm h−1 respectively. Dotted gray line represents the predictability threshold (1/e).

Download

Considering the advantages of transfer learning that can be effectively applied when data for new domains are insufficient over pre-trained domains, we evaluated the results of using pre-trained models that had been trained for additional dam basins: Daecheong Dam, Juam Dam, and Yongdam Dam basins in addition to Soyang-gang Dam basin (Cases 4 and 5). Through the R for 10 min precipitation prediction at dam basins, Cases 2 and 3, which used the pre-trained model with Soyang-gang Dam basin, show better performance for both Andong Dam and Chungju Dam (Table 5), but as the lead time increased to 90 min, Cases 4 and 5 maintain better performance for up to 90 min (Figs. 8a and 9a). This trend is notable in RMSE and NSE. In addition, at CSI values of higher rainfall intensity, Cases 4 and 5 outperform other strategies at the longer lead time for both Andong Dam and Chungju Dam basins (Figs. 8e–f and 9e–f). As various and numerous data can solve the problem of data imbalance that causes underestimation of the model, the CSI value is good even at high rainfall intensity (Wang et al., 2016). In addition, Fang et al. (2022) showed that models trained through diverse and numerous data on multiple regions can also learn about the characteristics that contribute to regional differences and are more effective in predicting extreme events and future trends. These results show that the diversity and amount of data have no significant effect on the short-term prediction of low rainfall intensity but are very important in resolving model underestimation and improving prediction accuracy for heavy rainfall.

4 Conclusions

In this study, our aim was to develop a precipitation nowcasting model for the dam basin so that it could be applied at each flood control center. We developed a model that could perform sufficiently well with the relatively simple structure and low computational costs and evaluated the applicability of the transfer learning technique to facilitate its application in multiple dam basins. The proposed model could be used for rainfall-runoff modeling in dam basins in future work. To develop the radar-based precipitation nowcasting model, we applied a cGAN approach based on the U-net architecture. The model architecture was inspired by the image-to-image translation model called Pix2Pix, which consists of U-Net as the generator and PatchGAN as the discriminator (Isola et al., 2017). In 10 min lead time precipitation prediction, at the Soyang-gang Dam basin, our model outperformed the other reference models. Additionally, when we applied the recursive process to predict precipitation with lead times of up to 90 min, our model achieved adequate performance (>1/e) for R with lead times of up to about 80 min, which was an improvement over ConvLSTM (up to 60 min) and U-Net (up to 60 min). Also, the CSI and FSS (at the intensity of 0.1 mm h−1) results for the entire domain revealed that compared with the reference models, our model generated precipitation prediction more accurately at the overall lead times. However, in the case of higher rainfall intensity, CSI and FSS showed that Rad-cGAN had relatively poor performance compared with the reference models (excluding U-net). Although our model tends to underestimate strong precipitation, the qualitative evaluation of the typhoon Soulik confirmed that our model can capture spatiotemporal change in the area of precipitation closest to the ground truth. In addition, based on the PSD results, our model can generate sharper and more realistic images than U-net by applying the cGAN approach. Furthermore, considering the edge effect, pySTEPS showed improved performance using an extended input domain compared with the original input domain. However, because our model also performed better using extended data, we conclude that Rad-cGAN is the most advanced precipitation nowcasting model that does not consider edge effects compared with other reference models.

To develop the precipitation nowcasting model for different dam basins (Andong Dam and Chungju Dam basins), we proposed different transfer learning strategies by using the previously trained model. Comparing the cases of using transfer learning (Cases 2–5) and the case of not using transfer learning (Case 1), the cases of using transfer learning generally showed better performance in both Andong Dam and Chungju Dam basins. However, from the results of Case 3, in which the performance is somewhat poor in the case of using fine-tuning, it is necessary to pay attention to the overfitting when applying the fine-tuning procedure. In addition, when the model trained with additional dam basins was used as a pre-trained model (Cases 4 and 5), the prediction performance was outperformed, especially at high rainfall intensity, and it was found that data diversity affected model generalization and underestimation.

We confirmed that the proposed precipitation nowcasting model demonstrated improved performance over conventional machine-learning-based models (U-Net and ConvLSTM) and showed that transfer learning strategies could be effectively applied to develop models for other dam domains in South Korea. However, there are remaining issues that must be considered to ensure the auditability of our model for real problems, such as predicting heavy precipitation events and flash flood forecasting. First, the tendency of the model to underestimate precipitation is a major issue. The decisive cause of this issue is data imbalance, as mentioned in general machine-learning tasks (Wang et al., 2016). To address this issue, further studies need to be conducted to improve the predictive performance of extreme precipitation events by extending the duration of training data and assigning weights to the extreme or other events. Additionally, adding information about domain characteristics, such as the digital elevation model and the land cover map is expected to improve the precipitation nowcasting model. Another issue is that we trained models for different domains using basic transfer learning strategies, and evaluated the performance only for the new domains, which are not sufficient to develop models for multiple dam domains that can be used in early warning systems. To overcome this issue, for example, Wang et al. (2020) presented a new transfer learning approach that simultaneously mined the knowledge of multiple pre-trained generators. Therefore, further research using more advanced transfer learning strategies is expected to help the development of precipitation nowcasting models with superior performance in different domains to increase practicality.

Code and data availability

Source code of the model architecture, the pre-trained model for Soyang-gang Dam basin, and example test data are available at Zenodo (https://doi.org/10.5281/zenodo.6650722; Choi and Kim, 2021a). The model usage and implementation codes of reference models (U-net; ConvLSTM; pySTEPS) are available at GitHub repository (https://github.com/SuyeonC/Rad-cGAN; last access: 22 July 2022) and are also accessible via Zenodo: https://doi.org/10.5281/zenodo.6880997 (Choi and Kim, 2021b).

The radar reflectivity composite data used for training and evaluating the model can be downloaded at the Korea Meteorological Administration data portal: https://data.kma.go.kr/data/rmt/rmtList.do?code=11&pgmNo=62 (KMA, 2022) or are available upon request to the KMA.

Author contributions

SC and YK designed the study, and SC performed the model development, simulations, and result analysis under the supervision of YK. SC wrote the original manuscript, and YK reviewed and edited the manuscript.

Competing interests

The contact author has declared that neither of the authors has any competing interests.

Disclaimer

Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Financial support

This study was supported by the Basic Science Research Program through the National Research Foundation of Korea, which was funded by the Ministry of Science, ICT & Future Planning (grant no. 2020R1A2C2007670), and the Technology Advancement Research Program through the Korea Agency for Infrastructure Technology Advancement (KAIA) grant funded by the Ministry of Land, Infrastructure and Transport (grant no. 22CTAP-C163540-02).

Review statement

This paper was edited by Charles Onyutha and reviewed by three anonymous referees.

References

Agrawal, S., Barrington, L., Bromberg, C., Burge, J., Gazen, C., and Hickey, J.: Machine learning for precipitation nowcasting from radar images, arXiv [preprint], https://doi.org/10.48550/arXiv.1912.12132, 2019. 

Ayzel, G., Heistermann, M., and Winterrath, T.: Optical flow models as an open benchmark for radar-based precipitation nowcasting (rainymotion v0.1), Geosci. Model Dev., 12, 1387–1402, https://doi.org/10.5194/gmd-12-1387-2019, 2019. 

Ayzel, G., Scheffer, T., and Heistermann, M.: RainNet v1.0: a convolutional neural network for radar-based precipitation nowcasting, Geosci. Model Dev., 13, 2631–2644, https://doi.org/10.5194/gmd-13-2631-2020, 2020. 

Basist, A., Bell, G. D., and Meentemeyer, V.: Statistical relationships between topography and precipitation patterns, J. climate, 7, 1305–1315, https://doi.org/10.1175/1520-0442(1994)007<1305:SRBTAP>2.0.CO;2, 1994. 

Berenguer, M., Surcel, M., Zawadzki, I., Xue, M., and Kong, F.: The Diurnal Cycle of Precipitation from Continental Radar Mosaics and Numerical Weather Prediction Models. Part II: Intercomparison among Numerical Models and with Nowcasting, Mon. Weather Rev., 140, 2689–2705, https://doi.org/10.1175/MWR-D-11-00181.1, 2012. 

Choi, S. and Kim, Y.: Rad-cGAN v1.0: Radar-based precipitation nowcasting model with conditional Generative Adversarial Networks for multiple dam domains, Zenodo [code], https://doi.org/10.5281/zenodo.6650722, 2021a. 

Choi, S. and Kim, Y.: SuyeonC/Rad-cGAN: (v1.0.0), Zenodo [code], https://doi.org/10.5281/zenodo.6880997, 2021b. 

Clark, A., Donahue, J., and Simonyan, K.: Adversarial video generation on complex datasets, arXiv [preprint], https://doi.org/10.48550/arXiv.1907.06571, 2019. 

Fang, K., Kifer, D., Lawson, K., Feng, D., and Shen, C.: The data synergy effects of time-series deep learning models in hydrology, Water Resour. Res., 58, e2021WR029583, https://doi.org/10.1029/2021WR029583, 2022. 

Foresti, L. and Seed, A.: The effect of flow and orography on the spatial distribution of the very short-term predictability of rainfall from composite radar images, Hydrol. Earth Syst. Sci., 18, 4671–4686, https://doi.org/10.5194/hess-18-4671-2014, 2014. 

Germann, U. and Zawadzki, I.: Scale-Dependence of the Predictability of Precipitation from Continental Radar Images. Part I: Description of the Methodology, Mon. Weather Rev., 130, 2859–2873, https://doi.org/10.1175/1520-0493(2002)130<2859:SDOTPO>2.0.CO;2, 2002. 

Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y.: Generative adversarial networks, in: Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, Canada, 8–13 December 2014, The MIT Press, 2672–2680, available at: https://arxiv.org/pdf/1406.2661.pdf (last access: 20 July 2022), 2014. 

He, K., Zhang, X., Ren, S., and Sun, J.: Deep Residual Learning for Image Recognition, in: CVPR 2016: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, Nevada, USA, 27–30 June 2016, IEEE, 770–778, https://doi.org/10.1109/CVPR.2016.90, 2016. 

Hochreiter, S. and Schmidhuber, J.: Long Short-Term Memory, Neural Comput., 9, 1735–1780, https://doi.org/10.1162/neco.1997.9.8.1735, 1997. 

Hwang, S., Yoon, J., Kang, N., and Lee, D.-R.: Development of flood forecasting system on city  mountains  small river area in Korea and assessment of forecast accuracy, Journal of Korea Water Resources Association, 53, 225–236, https://doi.org/10.3741/JKWRA.2020.53.3.225, 2020. 

Imhoff, R. O., Brauer, C. C., Overeem, A., Weerts, A. H., and Uijlenhoet, R.: Spatial and Temporal Evaluation of Radar Rainfall Nowcasting Techniques on 1,533 Events, Water Resour. Res., 56, e2019WR026723, https://doi.org/10.1029/2019WR026723, 2020. 

Ioffe, S. and Szegedy, C.: Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift, in: ICML'15: Proceedings of the 32nd International Conference on International Conference on Machine Learning, Lille, France, 6–11 July 2015, JMLR.org, 37, 448–456, available at: http://proceedings.mlr.press/v37/ioffe15.pdf (last access: 20 July 2022), 2015. 

Isola, P., Zhu, J., Zhou, T., and Efros, A. A.: Image-to-Image Translation with Conditional Adversarial Networks, in: CVPR 2017: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, Hawaii, USA, 21–26 July 2017, IEEE, 1125–1134, https://doi.org/10.1109/CVPR.2017.632, 2017. 

Jeong, C. H., Kim, W., Joo, W., Jang, D., and Yi, M. Y.: Enhancing the Encoding-Forecasting Model for Precipitation Nowcasting by Putting High Emphasis on the Latest Data of the Time Step, Atmosphere, 12, 261, https://doi.org/10.3390/atmos12020261, 2021. 

Kim, S., Hong, S., Joh, M., and Song, S.-K.: Deeprain: Convlstm network for precipitation prediction using multichannel radar data, in: Proceedings of the 7th International Workshop on Climate Informatics: CI 2017, Boulder, Colorado, USA, 20–22 September 2017, UCAR/NCAR – Library, 89–92, https://arxiv.org/pdf/1711.02316.pdf (last access: 20 July 2022), 2017. 

Kingma, D. P. and Ba, J.: Adam: A method for stochastic optimization, in: Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, California, USA, 7–9 May 2015, Conference Track Proceedings, https://arxiv.org/pdf/1412.6980.pdf (last access: 21 July 2022), 2015. 

KMA (Korea Meteorological Administration): Quality-controlled 1.5 km Constant Altitude Plan-Position Indicator (CAPPI) Radar Reflectiviy Composite map, KMA [data set], https://data.kma.go.kr/cmmn/main.do, last access: 22 July 2022. 

Krizhevsky, A., Sutskever, I., and Hinton, G. E.: ImageNet classification with deep convolutional neural networks, in: Proceedings of the 26th International Conference on Neural Information Processing Systems, Lake Tahoe, Nevada, USA, 3–8 December 2012, NeurIPS, 1097–1105, https://proceedings.neurips.cc/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf (last access: 21 July 2022), 2012. 

Kumar, A., Islam, T., Sekimoto, Y., Mattmann, C., and Wilson, B.: Convcast: An embedded convolutional LSTM based architecture for precipitation nowcasting using satellite data, Plos One, 15, e0230114, https://doi.org/10.1371/journal.pone.0230114, 2020. 

Long, J., Shelhamer, E., and Darrell, T.: Fully convolutional networks for semantic segmentation, in: CVPR 2015: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, Massachusetts, USA, 7–12 June 2015, IEEE, 3431–3440, https://doi.org/10.1109/CVPR.2015.7298965, 2015. 

Marshall, J. S. and Palmer, W. M. K.: THE DISTRIBUTION OF RAINDROPS WITH SIZE, J. Atmos. Sci., 5, 165–166, https://doi.org/10.1175/1520-0469(1948)005<0165:Tdorws>2.0.Co;2, 1948. 

McCuen, R. H., Knight, Z., and Cutter, A. G.: Evaluation of the Nash-Sutcliffe Efficiency Index, J. Hydrol. Eng., 11, 597–602, https://doi.org/10.1061/(ASCE)1084-0699(2006)11:6(597), 2006. 

Mirza, M. and Osindero, S.: Conditional generative adversarial nets, arXiv [preprint], https://doi.org/10.48550/arXiv.1411.1784, 2014. 

Mo, S., Cho, M., and Shin, J.: Freeze the discriminator: a simple baseline for fine-tuning gans, in: Proceedings of the Conference on Computer Vision and Pattern Recognition: AI for Content Creation Workshop, Online, 15 June 2020, AICCW, https://arxiv.org/pdf/2002.10964.pdf (last access: 21 July 2022), 2020. 

Moishin, M., Deo, R. C., Prasad, R., Raj, N., and Abdulla, S.: Designing Deep-Based Learning Flood Forecast Model With ConvLSTM Hybrid Algorithm, IEEE Access, 9, 50982–50993, https://doi.org/10.1109/ACCESS.2021.3065939, 2021. 

Moriasi, D. N., Arnold, J. G., Van Liew, M. W., Bingner, R. L., Harmel, R. D., and Veith, T. L.: Model evaluation guidelines for systematic quantification of accuracy in watershed simulations, T. ASABE, 50, 885–900, https://doi.org/10.13031/2013.23153, 2007. 

Pan, S. J. and Yang, Q.: A Survey on Transfer Learning, IEEE T. Knowl. Data En., 22, 1345–1359, https://doi.org/10.1109/TKDE.2009.191, 2010. 

Pierce, C., Seed, A., Ballard, S., Simonin, D., and Li, Z.: Nowcasting, in: Doppler Radar Observations – Weather Radar, Wind Profiler, Ionospheric Radar, and Other Advanced Applications, edited by: Bech, J. and Chau, J. L., InTechOpen., 97–142, https://doi.org/10.5772/39054, 2012. 

Poletti, M. L., Silvestro, F., Davolio, S., Pignone, F., and Rebora, N.: Using nowcasting technique and data assimilation in a meteorological model to improve very short range hydrological forecasts, Hydrol. Earth Syst. Sci., 23, 3823–3841, https://doi.org/10.5194/hess-23-3823-2019, 2019. 

Prudhomme, C. and Reed, D. W.: Relationships between extreme daily precipitation and topography in a mountainous region: a case study in Scotland. Int. J. Climatol., 18, 1439–1453, https://doi.org/10.1002/(SICI)1097-0088(19981115)18:13<1439::AID-JOC320>3.0.CO;2-7, 1998. 

Pulkkinen, S., Nerini, D., Pérez Hortal, A. A., Velasco-Forero, C., Seed, A., Germann, U., and Foresti, L.: Pysteps: an open-source Python library for probabilistic precipitation nowcasting (v1.0), Geosci. Model Dev., 12, 4185–4219, https://doi.org/10.5194/gmd-12-4185-2019, 2019. 

Ravuri, S., Lenc, K., Willson, M., Kangin, D., Lam, R., Mirowski, P., Fitzsimons, M., Athanassiadou, M., Kashem, S., Madge, S., Prudden, R., Mandhane, A., Clark, A., Brock, A., Simonyan, K., Hadsell, R., Robinson, N., Clancy, E., Arribas, A., and Mohamed, S.: Skilful precipitation nowcasting using deep generative models of radar, Nature, 597, 672–677, https://doi.org/10.1038/s41586-021-03854-z, 2021. 

Renzullo, L., Velasco-Forero, C., and Seed, A.: Blending radar, NWP and satellite data for real-time ensemble rainfall analysis: a scale-dependent method, Tech. Rep. EP174236, CSIRO, https://doi.org/10.4225/08/594eb78c96025, 2017. 

Ronneberger, O., Fischer, P., and Brox, T.: U-Net: Convolutional Networks for Biomedical Image Segmentation, in: Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015, edited by: Navab, N., Hornegger, J., Wells, W. M., and Frangi, A. F., Springer International Publishing, Cham, 234–241, https://doi.org/10.1007/978-3-319-24574-4_28, 2015. 

Rüttgers, M., Lee, S., Jeon, S., and You, D.: Prediction of a typhoon track using a generative adversarial network and satellite images, Sci. Rep.-UK, 9, 6057, https://doi.org/10.1038/s41598-019-42339-y, 2019. 

Seed, A. W.: A dynamic and spatial scaling approach to advection forecasting, J. Appl. Meteorol., 42, 381–388, https://doi.org/10.1175/1520-0450(2003)042<0381:ADASSA>2.0.CO;2, 2003. 

Shi, X., Chen, Z., Wang, H., Yeung, D.-Y., Wong, W.-K., and Woo, W.-C.: Convolutional LSTM Network: a machine learning approach for precipitation nowcasting, in: Proceedings of the 29th International Conference on Neural Information Processing Systems, Montreal, Canada, 7–12 December 2015, NeurIPS, 802–810, https://proceedings.neurips.cc/paper/2015/file/07563a3fe3bbe7e3ba84431ad9d055af-Paper.pdf (last access: 21 July 2022), 2015. 

Simonyan, K. and Zisserman, A.: Very deep convolutional networks for large-scale image recognition, in: Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, California, USA, 7–9 May 2015, Conference Track Proceedings, https://arxiv.org/pdf/1409.1556.pdf (last access: 21 July 2022), 2015. 

Sønderby, C. K., Espeholt, L., Heek, J., Dehghani, M., Oliver, A., Salimans, T., Agrawal, S., Hickey, J., and Kalchbrenner, N.: Metnet: A neural weather model for precipitation forecasting, arXiv [preprint], https://doi.org/10.48550/arXiv.2003.12140, 2020. 

Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., and Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting, J. Mach. Learn. Res., 15, 1929–1958, 2014. 

Sutskever, I., Vinyals, O., and Le, Q. V.: Sequence to sequence learning with neural networks, , in: Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, Canada, 8–13 December 2014, NeurIPS, 3014–3112, https://proceedings.neurips.cc/paper/2014/file/a14ac55a4f27472c5d894ec1c3c743d2-Paper.pdf (last access: 20 July 2022), 2014. 

Trebing, K., Staǹczyk, T., and Mehrkanoon, S.: SmaAt-UNet: Precipitation nowcasting using a small attention-UNet architecture, Pattern Recogn. Lett., 145, 178–186, https://doi.org/10.1016/j.patrec.2021.01.036, 2021.  

Wang, S., Liu, W., Wu, J., Cao, L., Meng, Q., and Kennedy, P. J.: Training deep neural networks on imbalanced data sets, in: 2016 International Joint Conference on Neural Networks (IJCNN), 24–29 July 2016, 4368–4374, https://doi.org/10.1109/IJCNN.2016.7727770, 2016. 

Wang, Y., Coning, E., Harou, A., Jacobs, W., Joe, P., Nikitina, L., Roberts, R., Wang, J., and Wilson, J.: Guidelines for nowcasting techniques, WMO publication, https://library.wmo.int/opac/doc_num.php?explnum_id=3795 (last access: 19 July 2022), 2017. 

Wang, Y., Wu, C., Herranz, L., van de Weijer, J., Gonzalez-Garcia, A., and Raducanu, B.: Transferring gans: generating images from limited data, in: Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018, Springer, Cham, 220–236, https://doi.org/10.1007/978-3-030-01231-1_14, 2018. 

Wang, Y., Gonzalez-Garcia, A., Berga, D., Herranz, L., Khan, F. S., and Weijer, J. v. d.: MineGAN: Effective Knowledge Transfer From GANs to Target Domains With Few Images, in: CVPR 2020: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Online, 14–19 June 2020, IEEE, 9332–9341, https://doi.org/10.1109/CVPR42600.2020.00935, 2020. 

Download
Short summary
Here we present the cGAN-based precipitation nowcasting model, named Rad-cGAN, trained to predict a radar reflectivity map with a lead time of 10 min. Rad-cGAN showed superior performance at a lead time of up to 90 min compared with the reference models. Furthermore, we demonstrate the successful implementation of the transfer learning strategies using pre-trained Rad-cGAN to develop the models for different dam domains.