the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Assessment of the data assimilation framework for the Rapid Refresh Forecast System v0.1 and impacts on forecasts of a convective storm case study
Ivette H. Banos
Will D. Mayfield
Guoqing Ge
Luiz F. Sapucci
Jacob R. Carley
Louisa Nance
Download
- Final revised paper (published on 12 Sep 2022)
- Preprint (discussion started on 05 Oct 2021)
Interactive discussion
Status: closed
-
RC1: 'Comment on gmd-2021-289', Anonymous Referee #1, 11 Nov 2021
General comments:
The paper is overall well structured, clearly described, and provides a succinct evaluation of a single use case used for tuning some parameters of RRFS. My one critical comment here is that the title may be a bit misleading, and should be modified if possible (something along the lines of ‘... on forecasts of a convective storm case study’. As it stands, readers are at first likely expecting a larger, more comprehensive, data assimilation evaluation paper consisting of multiple case studies and deeper analyses. To be clear, this single case study paper is useful, but the correct expectation should be set with the title.
Specific comments:
- GSI is capable of hybrid 4DEnVar. Is there a reason this flavor of DA was not included in the comparisons?
- For those not familiar with how rapid refresh systems are typically cycled, why is it necessary to perform a periodic cold start even though hourly DA is performed?
- You mention the great importance of tuning localization parameters but only vertical localization is tuned, why is it assumed that the default horizontal localization does not need tuning?
technical corrections
- At first I was confused by the different version numbers RRFSv1a/RRFSv0.1, perhaps it would be useful to clarify early on that these are the physics suite / cycling system
- 64 - define what the “convective gray zone” is for readers who might not be unfamiliar
- 92 - “as good” -> “as well”
Citation: https://doi.org/10.5194/gmd-2021-289-RC1 -
AC4: 'Reply on RC1', Ivette Hernandez Banos, 11 Feb 2022
General comments:
The paper is overall well structured, clearly described, and provides a succinct evaluation of a single use case used for tuning some parameters of RRFS. My one critical comment here is that the title may be a bit misleading, and should be modified if possible (something along the lines of ‘... on forecasts of a convective storm case study’. As it stands, readers are at first likely expecting a larger, more comprehensive, data assimilation evaluation paper consisting of multiple case studies and deeper analyses. To be clear, this single case study paper is useful, but the correct expectation should be set with the title.
We thank the reviewer for the comments and suggestions. Our responses are noted below. Changes to the document are highlighted in magenta.
The title was modified to “Assessment of the data assimilation framework for the Rapid Refresh Forecast System v0.1 and impacts on forecasts of a convective storm case study”
Specific comments:
- GSI is capable of hybrid 4DEnVar. Is there a reason this flavor of DA was not included in the comparisons?
The Rapid Refresh Forecast System (RRFS) is intended for hybrid 3DEnVar data assimilation following the currently operational Rapid Refresh (RAP) configuration. As in RAP, hourly updated cycles are configured in order to leverage available data with higher frequency such as surface observations. Testing and development of a hybrid 4DEnVar system is a worthwhile pursuit, but due to limited resources it is beyond the scope of this article describing the v0.1 capability.
- For those not familiar with how rapid refresh systems are typically cycled, why is it necessary to perform a periodic cold start even though hourly DA is performed?
Rapid refresh systems are usually configured for regional domains in which a reduced amount of data is used compared to global domains. Benjamin et al. (2016) pointed out the importance of a partial cycling technique to update large scale conditions such as the longwave representation. This is of greater importance in regional domains with coverage over oceans such as in RAP and thus in RRFS, where cycles with a warm start from a parallel partial cycle (currently only cold start in RRFS v0.1) are used in order to improve the drift from the global scale. A sentence was added to account for the reviewer's comment.
Lines 241-243: “Periodic updates of the large scale atmospheric conditions are needed in regional modeling systems in order to account for corrections made by global observations over land and ocean and to avoid model drift from those conditions (Benjamin et al., 2016).”
- You mention the great importance of tuning localization parameters but only vertical localization is tuned, why is it assumed that the default horizontal localization does not need tuning?
We agree with the reviewer. RRFS v1 will use a convective-scale ensemble and therefore the horizontal localization also needs to be tuned in order to assess if the default localization is good enough or if a different value leads to better results. In this study using RRFS v0.1, we use global ensemble members and seek to establish a measure of sensitivity to such parameters to inform further development. For convective systems the vertical representation is very important for storm initialization. Due to limited resources and time, and taking into account that a 3 km grid spacing is used in the experiments, similar to the High Resolution Rapid Refresh (HRRR) horizontal resolution, we decided to only present results of tuning the vertical localization. Testing and evaluating the horizontal localization in RRFS is underway in a separate work. A sentence was added to address the reviewer’s comment.
Lines 312-313: “A separate study is underway in which the optimal horizontal localization for RRFS is also investigated and therefore it is not examined here.”
Technical corrections:
- At first I was confused by the different version numbers RRFSv1a/RRFSv0.1, perhaps it would be useful to clarify early on that these are the physics suite / cycling system
In order to avoid confusion between these acronyms, RRFSv1a was changed to RRFS_PHYv1a and is more clearly written that RRFS v0.1 is the prototype RRFS used in this study. Lines 67, 111, 152-153, 594 were modified to account for these changes.
Line 66-67: “...a suite based on GFS version 16 physical parameterizations and a prototype of the RRFS physics suite (henceforth called RRFS_PHYv1a).”
Line 105: “ For the purpose of this paper, the prototype RRFS used is called RRFS v0.1”
- 64 - define what the “convective gray zone” is for readers who might not be unfamiliar
Line 65 was modified in order to provide a more straightforward argument of what was investigated in Harrold et al. (2021):
Line 65: “Harrold et al. (2021) investigated how the SRW represents convection and associated precipitation for varied model grid spacing in two physics suites…”
- 92 - “as good” -> “as well”
As suggested, the correction was made in line 94.
Line 94: “…it is imperative that the data assimilation component behave as well as or better than the current operational state-of-the-art”
References:
Benjamin, S. G., Weygandt, S. S., Brown, J. M., Hu, M., Alexander, C. R., Smirnova, T. G., Olson, J. B., James, E. P., Dowell, D. C.,Grell, G. A., et al.: A North American hourly assimilation and model forecast cycle: The Rapid Refresh, Monthly Weather Review, 144,1669–1694, https://journals.ametsoc.org/view/journals/mwre/144/4/mwr-d-15-0242.1.xml, 2016
Harrold, M., Hertneky, T., Kalina, E., Newman, K., Ketefian, G., Grell, E. D., Lybarger, N. D., and Nelson, B.: Investigating the Scalability ofConvective and Microphysics Parameterizations in the Unified Forecast System Short-Range Weather (UFS-SRW) Application, in: 101stAmerican Meteorological Society Annual Meeting, AMS, 2021
Citation: https://doi.org/10.5194/gmd-2021-289-AC4
-
RC2: 'Comment on gmd-2021-289', Anonymous Referee #2, 20 Nov 2021
The paper tested different GSI assimilation settings with FV3 LAM runs, and proved that the system can predict the severe convective squall line case over Oklahoma on 4 May, 2020. The results are useful to most NCEP forecast data users. The paper has some unclear or incomplete reasoning but will likely be a significant contribution with revision and clarification.
General comments:
For most experiments in this paper, only the comparison results were shown, and the specific causes of the results were not analyzed.
For example, the differences between PSEUDO and 75EnBEC experiments were huge, but the authors have not given many diagnostics. Why more observations through GSI will cause overestimated convection?
Another example, the VLOC of 1 layer should capture finer vertical features of low atmosphere but the result showed that the positive impact is above 650 hPa and negative impacts are below 800hpa, why?
If the RRFS aims to replace the NCEP operational suite of regional and convective scale modeling systems in the next upgrade, it would be best to show the result from RAP as a baseline for all these tests.
Specific comments:
In Figure 2. RRFS cycling configuration diagram, the cold start is at 0 utc and the warm start seems to be from 1 utc to 6 utc. But in Figure 5 and relative context, the cold start is at 0 utc and 12 utc, what is the exact cold start interval?
L127 LAM appeared first time, should be limited area modeling (LAM) capability
In Figure 9, no “Matched pair counts used for RMSE and bias computation at each cycle” were found in the photos.
Citation: https://doi.org/10.5194/gmd-2021-289-RC2 -
AC2: 'Reply on RC2', Ivette Hernandez Banos, 11 Feb 2022
The paper tested different GSI assimilation settings with FV3 LAM runs, and proved that the system can predict the severe convective squall line case over Oklahoma on 4 May, 2020. The results are useful to most NCEP forecast data users. The paper has some unclear or incomplete reasoning but will likely be a significant contribution with revision and clarification.
We thank the reviewer for the comments and suggestions. Our responses are noted below. Changes to the document are highlighted in red.
General comments:
For most experiments in this paper, only the comparison results were shown, and the specific causes of the results were not analyzed.
For example, the differences between PSEUDO and 75EnBEC experiments were huge, but the authors have not given many diagnostics. Why more observations through GSI will cause overestimated convection?
- The experiment PSEUDO needed to be rerun because of a bug in the calculation of the planetary boundary layer (PBL) height which affected the PBL pseudo-observations function. This bug only affected this experiment. Accordingly, figures 11, 12, and 13 were replaced with the new results and corresponding discussions and some diagnostics were provided in Sect. 4.3 (lines 512-537) in the new version of the manuscript, as suggested by the reviewer.
Another example, the VLOC of 1 layer should capture finer vertical features of low atmosphere but the result showed that the positive impact is above 650 hPa and negative impacts are below 800hpa, why?
- Indeed, the experiment VLOC was conducted in order to capture finer features of the low atmosphere through the reduction of the vertical localization from 3 levels to 1 level in the first 10 model layers. However, the results did not show a positive impact in the lower atmosphere but above 650 hPa. In part this is because the multivariate relationships within the background error covariance that spread the impact to different levels and locations. Also, because of the cycling technique, the impacts in the forecasts are found in other levels. This led us to conclude that the default value of 3 layers already gives the best results in most vertical levels and that the value of 1 layer may be too small considering that the distance between layers in the lower atmosphere is also small. A sentence was added to add more discussion to these results.
Lines 500-503: “The analysis cycling technique and multivariate relationships in the BEC spread the observations impact throughout different levels and locations, which could have led to the slight positive impact above 650 hPa instead of the lower atmosphere where the modification in the vertical localization was made.”
If the RRFS aims to replace the NCEP operational suite of regional and convective scale modeling systems in the next upgrade, it would be best to show the result from RAP as a baseline for all these tests.
- This study provides very preliminary results on how the GSI analysis system and the limited area capability of the Finite Volume Cubed Sphere dynamical core (FV3 LAM) performs with the different options tested. It provides an evaluation of different functions and parameter values used currently in HRRR and RAP. However, some functionalities are still being developed and tested and therefore a more comprehensive study with more up-to-date developments of RRFS is underway where a comparison against RAP (as well as HRRR, NAM, and NAM nests) will be shown.
Specific comments:
In Figure 2. RRFS cycling configuration diagram, the cold start is at 0 utc and the warm start seems to be from 1 utc to 6 utc. But in Figure 5 and relative context, the cold start is at 0 utc and 12 utc, what is the exact cold start interval?
- We agree that the diagram could mislead readers. It was modified to include cycles from 00:00 UTC to 12:00 UTC. Lines 247-2I 50 were modified accordingly.
Lines 247-250: “ Figure 2 illustrates the RRFS cycling configuration from cycles initialized between 00:00 UTC through 12:00 UTC. In each cycle, an 18 h free forecast is launched following the analysis, with hourly outputs. A cold start is performed at 00:00 UTC and 12:00 UTC and warm starts between 01:00 UTC to 11:00 UTC using the FV3 LAM 1 h forecast from the previous cycle as background for the analysis”
L127 LAM appeared first time, should be limited area modeling (LAM) capability
- We added the LAM definition right after its first appearance in Line 55, as suggested.
In Figure 9, no “Matched pair counts used for RMSE and bias computation at each cycle” were found in the photos.
- We thank the reviewer for pointing out this mistake. The sentence was removed from the caption of Figure 9.
Citation: https://doi.org/10.5194/gmd-2021-289-AC2
-
AC2: 'Reply on RC2', Ivette Hernandez Banos, 11 Feb 2022
-
RC3: 'Comment on gmd-2021-289', Anonymous Referee #3, 23 Nov 2021
General comments:
I appreciate the extended introduction and literature review of the state of the modeling and DA development for this application. The manuscript provides a useful snapshot with regards to the state of testing and evaluation of a case study using the RRFS during its development (v0.1).
For the benefit of the general data assimilation audience, some detail would be helpful to describe why the unconventional data assimilation cycle is used for the RRFS. For example, why is a cold start applied repeatedly throughout the DA cycles? Why can’t the RRFS be run as a continuous cycled process like conventional DA applications? This could perhaps be connected with an improved discussion of the FV3 LAM and its limitations over other possible regional modeling approaches.
Please be consistent with the tense throughout. Present tense is appropriate, but occasionally it slips into the use of past tense.
Specific comments:
L 4-5:
“The current data assimilation component uses the Gridpoint Statistical Interpolation (GSI) system.”
It would be helpful to mention here what DA method is being used. Is it 3D-Var?
L 6:
“Results show that a baseline RRFS run without data assimilation is able to represent the observed convection, but with stronger cells and large location errors.”
How does the RRFS represent observed convection without data assimilation? Is it through the boundary/initial conditions coming from the global model? In that case, it would be using data assimilation indirectly through the global analysis. Could the authors please clarify.
L 8-9:
“using 75 % of the ensemble background error covariance (BEC)”
What does it mean to use only 75% of the BEC?
L 9-10:
“Decreasing the vertical ensemble localization radius in the first 10 layers of the hybrid analysis results in overall less skillful forecasts.”
From what initial radius to what final radius?
Please change to:
“Decreasing the vertical ensemble localization radius [from X m to Y m] in the first 10 layers of the hybrid analysis results in overall less skillful forecasts.”
L 97-99:
“Using hybrid 3DEnVar with 75 % of the ensemble background error covariance (BEC) showed storm structures in the 2 h forecast comparable to when using ensemble Kalman (EnKF), although EnKF outperformed 3DEnVar in the first hour forecast.”
(1) Does this mean that the BEC is weighted 75% toward the dynamic ensemble-estimated BEC and 25% to the static climatological BEC?
(2) If the EnKF outperforms the 3DEnVar in the first hour, and then they are comparable in the second hour, then when not use the EnKF instead of the 3DEnVar?
L 99-100:
“Both methods showed higher equitable threat scores (ETS) when compared to 3DVar and pure ensemble during the 4 h forecast analyzed.”
Does “both methods” refer to the3DEnVar and the EnKF? What is a “pure ensemble”?
L 126-127:
“The FV3, originally a global model, features three types of local refinement capabilities: stretching of the global grid (Harris et al., 2016), nesting within the global grid (Harris and Lin, 2013), and a LAM capability (Black et al., 2021).”
It would be useful to mention briefly how these types of local refinement differ.
L 143-144:
“Hence, the CCPP contains a set of physical schemes and a common framework that facilitates the interaction between the physics and a numerical model (Bernardet et al., 2020). ”
Perhaps it would be more clear to say “between the physics parameterizations and the dynamical core”.
L 166:
“3D[V]ar”
L 171:
“3DEn[V]ar”
L 177-178:
Some discussion should be given about what deficiencies this will have. For example, the lower-resolution global ensemble members (which are even lower resolution than the global deterministic forecast) may have significant biases, and will not resolve the error characteristics at the scale of the LAM. Cleary, the global ensemble statistics provides some useful information, but it is not ideal.
L 223:
“cold starts are performed every 12 hours and warm starts are performed at all other cycles using the 1 h forecast from the previous cycle as background for the analysis.”
Why is the DA continually reset with cold and warm starts? What prevents the standard self-contained forecast-analysis cycle? Is there a model drift when the DA is run continually in the RRFS?
Figure 2:
It would help to have more annotation here. Where are the 18h forecasts coming from?
L 259:
“aircrafts”
Change to:
“aircraft”
L 263-264:
“The time window used is 1 hour, allowing for observations within 30 minutes before to 30 minutes after the analysis time to be assimilated.”
I suppose this implies that the analysis time is at the middle of the 1-hour forecast window, but I didn’t notice this mentioned earlier.
L 265-267:
“ For the hybrid 3DEnVar analysis, the Global Data Assimilation System (GDAS) 80 member ensemble forecasts (9 h forecasts) are used to provide the ensemble BEC (e.g. Wu et al., 2017).”
Please mention the resolution of these ensemble members. How well do the lower resolution members resolve dynamics at the scale within the 3km resolution LAM? E.g. how many grid points of the low-res FV3 global ensemble member fall within the LAM region?
L 268-269:
“For example, the 9 h GDAS ensemble forecasts initialized at 00:00 UTC (valid at 09:00 UTC) are used for the cycles from 07:00 UTC to 12:00 UTC.”
How is the 9-hour forecast initialized at 0 UTC used at 12 UTC? Is this using the BEC fixed in time at 9 UTC for the entire window?
L 271-272:
“ In all experiments with data assimilation, two outer loops with 50 inner loops each are performed to minimize the cost function and find each analysis.”
I’m not sure I understand what is done in the outer and inner loops. Are the inner loops referring to the PCG solver? If so, then was is done in the outer loop?
L 274-275:
“for different [applications].”
L 277-279:
“This baseline experiment is called NoDA and uses the same cycling configuration as experiments with data assimilation but without the execution of GSI.”
I understand that from your perspective this doesn’t use the GSI, but for the entire procedure the GSI is used at the global scale. I think it would be helpful to explain this context in more detail.
L 281:
“experiments with different ensemble weights [are] conducted”
L 335:
“The analysis residuals (OmA) are also depicted in Fig. 4”
The OmA’s are less useful than the OmF’s for assessing the DA performance. The analysis can be drawn arbitrarily close to the observations (e.g. using complete replacement). It is more valuable to see that the forecasts are being drawn closer to the future observations.
L 343:
“while some MESONET observations have large analysis residuals.”
What is the expected cause of this larger discrepancy in these observations?
L 350:
“This means the analyses fit closer to the observations, which is expected from a correctly executed data assimilation procedure.”
This is partially true, but it is also not correct to fit the observations perfectly (due to observational error), so care should be taken in such a statement.
L 351-353:
“There is a noticeable jump in the RMS error values of the OmB from 00:00 UTC (12:00 UTC) to 01:00 UTC (13:00 UTC) on 4 May 2020. This is because 00:00 UTC and 12:00 UTC are cold started from HRRR analyses. ”
So why are you using the cold starts?
Figure 5:
It would be useful to add a panel showing the difference between 3dvar and the 75EnVar.
L 365:
“the EnKF filter method.”
Change to:
“the EnKF method.”
L 377:
“Figure 7 shows the 2, 4, and 6 h forecasts of …”
The analysis of these results seems very subjective - could the authors please provide some general statistics for each experiment result to help in the comparison, and provide more objectivity. Below, for example, I suggest adding RMSE statistics to each sub-plot in Figure 7.
Figure 6:
Where does the error covariance matrix come from for the 3D-Var? How was it computed?
L 405-407:
“The wind RMSE results do not clearly indicate which experiment is best, but in general 100EnBEC shows the lowest values when considering all vertical levels.”
The hybrid methods generally underperform when the static BEC is inadequate. How is the BEC computed for these experiments? This may indicate that tuning is necessary for the B matrix in the hybrid, e.g. see:
Chang et al., 2020:
https://journals.ametsoc.org/view/journals/mwre/148/6/mwrD190128.xml
Also, the online estimation of the hybrid weighting parameter has been explored by De Azevedo et al. 2020, which may be worth mentioning:
https://www.tandfonline.com/doi/full/10.1080/16000870.2020.1835310
Figure 7:
It would be helpful if each one of these plots had an RMSE value appended (e.g. in the lower left corner) to make it easier to compare the methods.
Figure 8:
The Green RMSE is difficult to read. A slightly darker color would help. Also, perhaps you could order the RMSE in each plot from lowest to highest so that is it easier to see how each method performs in comparison to the others.
L 430:
“this study looked at”
Change to:
“this study looks at”
Figure 8 and 10:
I’m not sure that I see the value of the confidence interval shading. However are these confidence intervals computed? Can the authors justify that this statistic is meaningful for this application? (e.g., the confidence interval implies that this specific case can be extrapolated to all other relevant storm instances - perhaps more discussion of the computation and application of this method would be warranted.)
L 450-451:
“The impact of adding PBL pseudo-observations to the analysis based on surface temperature and moisture observations is evaluated in experiment PSEUDO.”
I may have missed it, but I don’t see the source of the pseudo-observations. What measurements are being converted to these pseudo-ops? Is there a reason the source measurements cannot be assimilating using an appropriate observation operator?
L 477:
“More tuning and testing of this function are needed before applying this technique in the RRFS.”
The results (e.g. in Figure 12) seem to indicate the use of pseudo-observations may be a bad idea. Is there any reason why continued effort should be made to do this?
Figure 13:
Based on Figure 12, this seems to be over-fitting the data, which is then causing problems throughout the model column. It seems like this result would be better achieved via post processing, so that it doesn’t have negative ripple effects throughout the cycled DA.
Figure 14:
Please state what the units and physical height correspond to model level 50.
L 559:
“Supersaturation clipping in GSI can improve specific humidity fields in the analyses, allowing for more realistic storm and precipitation forecasts at longer forecast lengths. At shorter forecast lead hours, it produces more spurious convection”
This sounds like the wrong modification is being made to correct a longer-term bias. It would be beneficial to track down the root cause that accumulates to cause the longer-term bias without degrading the short-term skill.
L 572:
“More extensive testing of RRFS, covering a wider variety of cases, larger domain, and longer period of time, is needed to demonstrate whether results found here are robust or may be case dependent.”
I agree, which is why I’m confused by the presentation of the confidence intervals. Could the authors either remove these in the plots above or describe in greater detail how they were computed and why they are relevant in this case.
Citation: https://doi.org/10.5194/gmd-2021-289-RC3 -
AC3: 'Reply on RC3', Ivette Hernandez Banos, 11 Feb 2022
General comments:
I appreciate the extended introduction and literature review of the state of the modeling and DA development for this application. The manuscript provides a useful snapshot with regards to the state of testing and evaluation of a case study using the RRFS during its development (v0.1).
For the benefit of the general data assimilation audience, some detail would be helpful to describe why the unconventional data assimilation cycle is used for the RRFS. For example, why is a cold start applied repeatedly throughout the DA cycles? Why can’t the RRFS be run as a continuous cycled process like conventional DA applications? This could perhaps be connected with an improved discussion of the FV3 LAM and its limitations over other possible regional modeling approaches.
Please be consistent with the tense throughout. Present tense is appropriate, but occasionally it slips into the use of past tense.
We thank the reviewer for the comments and suggestions, which were very useful to improve the presentation and scientific quality of the results discussion of this manuscript. Our responses are noted below. Changes to the document are highlighted in blue.
This study represents an evaluation of the initial operating capability of the RRFS. Subsequent development underway includes establishing a partial cycling capability for the inaugural operational implementation, version 1, with subsequent plans to consider a fully cycled version in later versions leveraging recent advances discussed in Schwartz et al. (2021). Lines 244 -247 were modified to include this comment.
Lines 244-247: “Development currently underway includes establishing a partial cycling capability for the inaugural operational implementation, RRFS version 1, with subsequent plans to consider a fully cycled version in later versions leveraging recent advances discussed in Schwartz et al. (2021).”
The verbal tense was corrected throughout the manuscript.
Specific comments:
L 4-5: “The current data assimilation component uses the Gridpoint Statistical Interpolation (GSI) system.”
It would be helpful to mention here what DA method is being used. Is it 3D-Var?
- As suggested, the text was modified to include the data assimilation method used, the hybrid three dimensional ensemble–variational data assimilation (3DEnVar):
Lines 4-5: “The current data assimilation component uses the hybrid three-dimensional ensemble-variational data assimilation (3DEnVar) algorithm in the Gridpoint Statistical Interpolation (GSI) system.”
L 6: “Results show that a baseline RRFS run without data assimilation is able to represent the observed convection, but with stronger cells and large location errors.”
How does the RRFS represent observed convection without data assimilation? Is it through the boundary/initial conditions coming from the global model? In that case, it would be using data assimilation indirectly through the global analysis. Could the authors please clarify.
- As we describe in Sect. 3.2 and 3.3, the boundary/initial conditions used in this study are from the 3 km grid High Resolution Rapid Refresh (HRRR) and the experiment without data assimilation uses the same cycling configuration as experiments with data assimilation, in terms of the cold and warm starts. Therefore, it would use data assimilation indirectly through the lateral boundary conditions at each analysis hour and through the cold start initial conditions at 00:00 UTC and 12:00 UTC. Two sentences were added in the abstract and in Sect. 3.3 to clarify this point.
Line 9: “... without data assimilation...” was removed.
Lines 7-9: “A domain of 3 km horizontal grid-spacing is configured and hourly update cycles are performed using initial and lateral boundary conditions from the 3 km grid High Resolution Rapid Refresh (HRRR).”
Lines 299-301: “The 3 km ICs from the HRRR are consistent with the 3 km grid-spacing of the RRFS, such that fine scale features found in the HRRR are present in the RRFS ICs.”
L 8-9: “using 75 % of the ensemble background error covariance (BEC)”
What does it mean to use only 75% of the BEC?
- The information provided was incomplete, the lines were modified to include the weight given to the ensemble and static parts of the background error covariance:
Lines 10-11:“...especially in the 4 and 6 h forecasts using 75 % of the ensemble background error covariance (BEC) and 25 % of the static BEC…”
L 9-10: “Decreasing the vertical ensemble localization radius in the first 10 layers of the hybrid analysis results in overall less skillful forecasts.”
From what initial radius to what final radius?
Please change to:
“Decreasing the vertical ensemble localization radius [from X m to Y m] in the first 10 layers of the hybrid analysis results in overall less skillful forecasts.”
- The suggested change was adopted:
Lines 12-13: “Decreasing the vertical ensemble localization radius from 3 layers to 1 layer in the first 10 layers of the hybrid analysis results in overall less skillful forecasts.”
L 97-99: “Using hybrid 3DEnVar with 75 % of the ensemble background error covariance (BEC) showed storm structures in the 2 h forecast comparable to when using ensemble Kalman (EnKF), although EnKF outperformed 3DEnVar in the first hour forecast.”
(1) Does this mean that the BEC is weighted 75% toward the dynamic ensemble-estimated BEC and 25% to the static climatological BEC?
(2) If the EnKF outperforms the 3DEnVar in the first hour, and then they are comparable in the second hour, then when not use the EnKF instead of the 3DEnVar?
- (1) Yes, the text was improved including the weight given to the static BEC:
Line 100: “... and 25 % of the static BEC…”
- (2) The reviewer would need to follow up with Tong et al. (2020), who conducted the work summarized in this portion of the literature review, to have this question addressed. For RRFS, development is underway to incorporate the EnKF into the hybrid data assimilation system for the first implementation.
L 99-100: “Both methods showed higher equitable threat scores (ETS) when compared to 3DVar and pure ensemble during the 4 h forecast analyzed.”
Does “both methods” refer to the 3DEnVar and the EnKF? What is a “pure ensemble”?
- Yes, “both methods” refer to the 3DEnVar and the EnKF. On the other hand, pure ensemble refers to pure En3DVar in which the BEC is composed of 100 % of the ensemble BEC and 0 % of the static BEC. This was clarified in the text:
Lines 101-103: “Both methods, hybrid En3DVar and EnKF, showed higher equitable threat scores (ETS) when compared to 3DVar and pure 3DEnVar during the 4 h forecast analyzed.”
L 126-127: “The FV3, originally a global model, features three types of local refinement capabilities: stretching of the global grid (Harris et al., 2016), nesting within the global grid (Harris and Lin, 2013), and a LAM capability (Black et al., 2021).”
It would be useful to mention briefly how these types of local refinement differ.
- The sentence has been updated for additional clarity.
Lines 129-132: “The FV3, originally a global model, features three types of local refinement capabilities: stretching of the global grid using the Schmidt refinement technique (Harris et al., 2016), one- and two-way nesting within the global grid (Harris and Lin, 2013), and recently a LAM capability (Black et al., 2021). The LAM capability eliminates the need to run a concurrent global model and instead relies upon lateral boundary conditions provided at pre-specified intervals from an external source.”
L 143-144: “Hence, the CCPP contains a set of physical schemes and a common framework that facilitates the interaction between the physics and a numerical model (Bernardet et al., 2020). ”
Perhaps it would be more clear to say “between the physics parameterizations and the dynamical core”.
- As suggested, we modified line 148 in the new version of the manuscript.
L 166: “3D[V]ar”
- The typo was corrected in line 170.
L 171: “3DEn[V]ar”
- The typo was corrected in lines 175 and 180.
L 177-178: Some discussion should be given about what deficiencies this will have. For example, the lower-resolution global ensemble members (which are even lower resolution than the global deterministic forecast) may have significant biases, and will not resolve the error characteristics at the scale of the LAM. Clearly, the global ensemble statistics provides some useful information, but it is not ideal.
- As suggested, some discussion was provided in lines 181-184 and lines 185-191.
Lines 181-184: “Although the use of lower-resolution global ensemble members may not be ideal for the representation of the error characteristics at finer scales, Wu et al. (2017) showed that considerable forecast improvement can be obtained even if the ensemble provided is from a different system, which is consistent with findings in other studies such as Hu et al., (2017).”
Lines 185-191: “Future work on RRFS involves the extension to a convective-scale ensemble in the EnKF, which will improve the representativeness associated with the forecast error covariance at finer scales. However, such a change is not a panacea. Aside from increased computational expense, the problem of rank deficiency of the ensemble-derived error covariance becomes more apparent with the expanded degrees of freedom associated with the finer spatial resolution. While localization helps somewhat, a computationally affordable ensemble is one that is often insufficiently sized. Therefore future work also includes efforts to introduce multiscale data assimilation capabilities, such as scale dependent localization (e.g., Huang et al., 2021).”
L 223: “cold starts are performed every 12 hours and warm starts are performed at all other cycles using the 1 h forecast from the previous cycle as background for the analysis.”
Why is the DA continually reset with cold and warm starts? What prevents the standard self-contained forecast-analysis cycle? Is there a model drift when the DA is run continually in the RRFS?
- The reviewer is right. Partial cycles are performed in order to avoid model drift when the data assimilation is run continually in the RRFS. Nevertheless, other methods are under consideration for future versions of RRFS, such as a fully cycled method with blending of longwaves from the global based on Schwartz et al. (2022). Lines 241-243 were modified for clarity.
Lines 241-243: “Periodic updates of the large scale atmospheric conditions are needed in regional modeling systems in order to account for corrections made by global observations over land and ocean and to avoid model drift from those conditions (Benjamin et al., 2016).”
Figure 2: It would help to have more annotation here. Where are the 18h forecasts coming from?
- The diagram of Figure 2 was improved showing more cycles in order to clarify that cold starts are performed at 00:00 UTC and 12:00 UTC. The acronym FV3 LAM was added to indicate that the 18 h forecasts are obtained as part of the model execution after the data assimilation.
L 259:
“aircrafts”
Change to:
“aircraft”
- The typo was corrected in line 276.
L 263-264: “The time window used is 1 hour, allowing for observations within 30 minutes before to 30 minutes after the analysis time to be assimilated.”
I suppose this implies that the analysis time is at the middle of the 1-hour forecast window, but I didn’t notice this mentioned earlier.
- Yes, the time window is centered at the analysis hour. Line 281 includes the word “central” to clarify this information.
L 265-267: “ For the hybrid 3DEnVar analysis, the Global Data Assimilation System (GDAS) 80 member ensemble forecasts (9 h forecasts) are used to provide the ensemble BEC (e.g. Wu et al., 2017).”
Please mention the resolution of these ensemble members. How well do the lower resolution members resolve dynamics at the scale within the 3km resolution LAM? E.g. how many grid points of the low-res FV3 global ensemble member fall within the LAM region?
- The resolution of the 80 member ensemble forecast is approximately 25 km. As mentioned in a previous comment, the use of lower-resolution global ensemble members may not be ideal for the representation of the fine scale motion seen at 3 km. However, doing so has shown to be beneficial (Hu et al. 2017) for the forecasts. It should be taken into account that this study uses a prototype RRFS, RRFS v0.1 and that future work already intends to address this issue. This was added in lines 181-184 and 288-289. Line 284 was also modified to include the resolution of the ensemble members, as suggested.
Lines 288-289: “As shown in Hu et al. (2017), using off-time global and fixed ensemble based BEC still produces better results than just using the static BEC.”
Line 284: “These forecasts have a horizontal resolution of approximately 25 km and…”
L 268-269: “For example, the 9 h GDAS ensemble forecasts initialized at 00:00 UTC (valid at 09:00 UTC) are used for the cycles from 07:00 UTC to 12:00 UTC.”
How is the 9-hour forecast initialized at 0 UTC used at 12 UTC? Is this using the BEC fixed in time at 9 UTC for the entire window?
- Yes, since GDAS ensemble forecasts are available four times per day, the same ensemble based BEC is used along the 6 h window. This is possible because GSI has the flexibility to use off-time ensemble forecasts, i.e. ensemble forecasts that do not match the analysis hour. As shown in Hu et al. (2017), using off-time global and fixed ensemble forecasts still produces better results than just using the static BEC. The following sentence was added to the manuscript:
Lines 288-289: “As shown in Hu et al. (2017), using off-time global and fixed ensemble based BEC still produces better results than just using the static BEC.”
L 271-272: “In all experiments with data assimilation, two outer loops with 50 inner loops each are performed to minimize the cost function and find each analysis.”
I’m not sure I understand what is done in the outer and inner loops. Are the inner loops referring to the PCG solver? If so, then was is done in the outer loop?
- In this study, the minimization was performed in two outer loops with 50 iterations each, which were sufficient to reach the convergence condition. In each outer loop a re-linearization is performed after the 50 iterations. The increment is zero for the first outer loop while for the second it is updated with the solution found after the 50 iterations of the first outer loop.
Lines 289-292: “In all experiments with data assimilation, two outer loops with 50 iterations each loop are performed to minimize the cost function and find each analysis. In each outer loop a re-linearization is performed (e.g., Kleist et al., 2009). The increment is zero for the first outer loop while for the second it is updated with the solution found after the 50 iterations of the first outer loop.“
L 274-275: “for different [applications].”
- As suggested, “practices” was changed to “applications” in line 296 of the new manuscript.
L 277-279: “This baseline experiment is called NoDA and uses the same cycling configuration as experiments with data assimilation but without the execution of GSI.”
I understand that from your perspective this doesn’t use the GSI, but for the entire procedure the GSI is used at the global scale. I think it would be helpful to explain this context in more detail.
- The sentence was modified in line 299 to clarify that the cycling configuration of the NoDA experiment follows the same configuration of the experiment with data assimilation, in terms of the cold and warm initial conditions. Explanation of what is performed in each initialization type is already provided in Sect. 2.7.
L 281: “experiments with different ensemble weights [are] conducted”
- The verbal tense was corrected in line 303 of the new manuscript.
L 335: “The analysis residuals (OmA) are also depicted in Fig. 4”
The OmA’s are less useful than the OmF’s for assessing the DA performance. The analysis can be drawn arbitrarily close to the observations (e.g. using complete replacement). It is more valuable to see that the forecasts are being drawn closer to the future observations.
- The reviewer is right. However, for a more complete assessment, results from both, OmA and OmF, are provided in Figure 5 for all cycles. Figure 4 intends to show the spatial distribution of the assimilated observation and inform readers on the performance of the data assimilation system.
L 343: “while some MESONET observations have large analysis residuals.”
What is the expected cause of this larger discrepancy in these observations?
- As pointed out in Morris et al., (2020), while some MESONET stations are well maintained, the majority do not meet siting standards and maintenance protocols. Therefore, larger residuals are expected from these observations when compared to other observation networks such as METAR, which is an officially maintained observation network. Owing to their uncertain quality, MESONET observations are therefore assigned a higher observation error via a station blacklist, which also explains the larger residuals. Lines 373-376 were modified to include more details.
Lines 373-376: “As pointed out in Morris et al., (2020), while some MESONET stations are well maintained, the majority do not meet siting standards and maintenance protocols and therefore are assigned a higher observation error via a station blacklist. As expected, larger residuals are found from these observations when compared to other observation networks.”
L 350: “This means the analyses fit closer to the observations, which is expected from a correctly executed data assimilation procedure.”
This is partially true, but it is also not correct to fit the observations perfectly (due to observational error), so care should be taken in such a statement.
- We agree with the reviewer. The sentence was modified to:
Lines 383-384: “This means the analyses fit the observations more closely, though owing to observation error not perfectly, which is expected from a correctly executed data assimilation procedure.”
L 351-353: “There is a noticeable jump in the RMS error values of the OmB from 00:00 UTC (12:00 UTC) to 01:00 UTC (13:00 UTC) on 4 May 2020. This is because 00:00 UTC and 12:00 UTC are cold started from HRRR analyses. ”
So why are you using the cold starts?
- As explained before, the cold starts in this study are basically to update the large-scale conditions and to correct the drift from the model. A more adequate cycling technique is being investigated as stated in lines 601-602 in the manuscript.
Figure 5: It would be useful to add a panel showing the difference between 3dvar and the 75EnVar.
- In the discussion provided for Figure 5 it is highlighted that the background root mean square error increases from 14:00 UTC to 23:00 UTC, and that in Figure 5b this increase is larger than in Figure 5a (3DVar vs. 75EnBEC). Therefore, arrows were added to Figure 5 to draw attention to that marked increase, which is the focus of this result. The differences for other periods of time are subtle between these two panels. Accordingly, the caption of Figure 5 was updated.
Caption: “Black arrows highlight the time period from 14:00 UTC to 23:00 UTC.”
L 365: “the EnKF filter method.”
Change to:
“the EnKF method.”
- The word “filter” was removed from line 398 in the new version of the manuscript.
L 377: “Figure 7 shows the 2, 4, and 6 h forecasts of …”
The analysis of these results seems very subjective - could the authors please provide some general statistics for each experiment result to help in the comparison, and provide more objectivity. Below, for example, I suggest adding RMSE statistics to each sub-plot in Figure 7.
- The reviewer is right. In the new version of the manuscript, the Method for Object-Based Diagnostic Evaluation (MODE) in the Model Evaluation Tools (MET) (Jensen et al., 2020) was used for reflectivity values larger than 35 dBZ. The median of maximum interest (MMI (F+O)) (Davis et al., 2009) derived from MODE results were then added to each subplot. This metric is an appropriate metric for this kind of verification, providing more objectivity to the discussion, as suggested by the reviewer. Lines 352-359 were added to the document in which the method and metric used are described. Corresponding discussion of these values was added to the new version of the manuscript in figures 7, 11, 15, and A1.
Lines 353-359: “Hourly MRMS composite reflectivity mosaics (optimal method) observations (Zhang et al., 2016) are used to verify the composite reflectivity forecasts using the Method for Object-Based Diagnostic Evaluation (MODE) in MET. In order to quantitatively identify the experiment configuration that yielded better forecasts, the median of maximum interest (MMI (F+O)) (Davis et al., 2009) is analyzed. This metric results from the median between the maximum interest from each observed object with all predicted objects (MIF), and the maximum interest from each predicted object with all observed objects (MIO). It takes into account all attributes used in the total interest calculation, summarizing them into a single value. The forecast in greatest agreement with the observations will give MMI (F+O) values closer to one. Otherwise, the values will be closer to zero.”
Figure 6: Where does the error covariance matrix come from for the 3D-Var? How was it computed?
- The static BEC for the 3DVar is the same currently used in RAP and HRRR. It is based on statistics from the Global Forecast System (GFS) and the North American Mesoscale Forecast System (NAM) forecasts. Line 306 was added to the new version of the manuscript.
Line 306: “The static BEC for the 3DVar is the same as currently used in RAP and HRRR (Benjamin et al. 2016).”
L 405-407: “The wind RMSE results do not clearly indicate which experiment is best, but in general 100EnBEC shows the lowest values when considering all vertical levels.”
The hybrid methods generally underperform when the static BEC is inadequate. How is the BEC computed for these experiments? This may indicate that tuning is necessary for the B matrix in the hybrid, e.g. see:
Chang et al., 2020:
https://journals.ametsoc.org/view/journals/mwre/148/6/mwrD190128.xml
- We agree with the reviewer, the static BEC matrix used may not be optimal for RRFS and efforts are underway in order to obtain this matrix using its own RRFS forecasts. As mentioned above, the static B used in the experiments is the one currently used in RAP and HRRR.
Line 450-451: “These results may indicate that the static BEC matrix used may not be optimal for RRFS v0.1 and efforts are underway in order to obtain a better BEC matrix.”
Also, the online estimation of the hybrid weighting parameter has been explored by De Azevedo et al. 2020, which may be worth mentioning:
https://www.tandfonline.com/doi/full/10.1080/16000870.2020.1835310
- As suggested, the work of Azevedo et al. (2020) was mentioned in lines 451-453.
Lines 451-453: “Moreover, an online estimation approach may be explored for the specification of the hybrid weighting parameter, such as the method proposed by Azevedo et al. (2020) in which a geographically varying weighting factor alpha is defined and the ensemble spread is used for the assignment of the weights.”
Figure 7: It would be helpful if each one of these plots had an RMSE value appended (e.g. in the lower left corner) to make it easier to compare the methods.
- As mentioned previously, in the new version of the manuscript, the Method for Object-Based Diagnostic Evaluation (MODE) in the Model Evaluation Tools (MET) (Jensen et al., 2020) was used for reflectivity values larger than 35 dBZ. The median of maximum interest (MMI (F+O)) (Davis et al., 2009) derived from MODE results were then added to each subplot. This metric provides more objectivity to the discussion, as suggested by the reviewer. Lines 353-359 were added to the document in which the method and metric used are described. Corresponding discussion of these values was added to the new version of the manuscript in figures 7, 11, 15, and A1.
Figure 8: The Green RMSE is difficult to read. A slightly darker color would help. Also, perhaps you could order the RMSE in each plot from lowest to highest so that is it easier to see how each method performs in comparison to the others.
- As suggested, Figure 8 was improved using a darker green for the 3DVar experiment results and the RMSE values were ordered from lowest to highest. Figures 9 and 16 were also modified to account for the color change in the 3DVar plot line.
L 430: “this study looked at”
Change to:
“this study looks at”
- The verbal tense was corrected in line 476 of the new version of the manuscript.
Figure 8 and 10: I’m not sure that I see the value of the confidence interval shading. However are these confidence intervals computed? Can the authors justify that this statistic is meaningful for this application? (e.g., the confidence interval implies that this specific case can be extrapolated to all other relevant storm instances - perhaps more discussion of the computation and application of this method would be warranted.)
- We agree with the reviewer that the confidence intervals interpretation was not added to the discussion and therefore it was improved in the new version of the manuscript. The confidence intervals used in this study were derived using a bootstrap resampling technique of 1000 replications with replacement at each forecast lead hour in every cycle, and with bias-corrected percentiles, as reported by Wilks, 2006 and Gilleland et al. (2018). The confidence intervals help to highlight where the differences between the experiments analyzed are significant or not. Adding the confidence intervals shading shows the variability in the sample and informs readers how confident we can be that the mean values plotted are actually different. However, since we are analyzing multiple forecasts for this one single case study, we are not confident that the results obtained in this study can be extrapolated to all other relevant storm instances. Nevertheless, some of the results show statistically significant differences for the sample of this case study, which were highlighted in the discussions of figures 8, 10, and 12 in the new version of the manuscript. As suggested, several lines in the document (435-444, 487-491, 526-529) were modified for clarity and to address the point made by the reviewer.
L 450-451: “The impact of adding PBL pseudo-observations to the analysis based on surface temperature and moisture observations is evaluated in experiment PSEUDO.”
I may have missed it, but I don’t see the source of the pseudo-observations. What measurements are being converted to these pseudo-ops? Is there a reason the source measurements cannot be assimilating using an appropriate observation operator?
- This function is currently used in operational RAP and HRRR. In Sect. 3.3, we provided readers a background on why this function was developed and referenced the studies of Benjamin et al., 2004, Smith et al. (2007), Benjamin et al. (2010), and Benjamin et al. (2016) where more details on this function can be found. For clarity, line 507 was updated.
Line 507: “...using METAR observations,…”
L 477: “More tuning and testing of this function are needed before applying this technique in the RRFS.”
The results (e.g. in Figure 12) seem to indicate the use of pseudo-observations may be a bad idea. Is there any reason why continued effort should be made to do this?.
- The experiment PSEUDO needed to be rerun because of a bug in the calculation of the planetary boundary layer (PBL) height which affected the PBL pseudo-observations function. This bug only affected this experiment. Accordingly, figures 11, 12, and 13 were replaced with the new results, which indicate that this strategy improves the convection in particular for the 4 h forecast. The new analysis of these results corresponding discussions were provided in Sect. 4.3 (lines 512-537) in the new version of the manuscript.
Figure 13: Based on Figure 12, this seems to be over-fitting the data, which is then causing problems throughout the model column. It seems like this result would be better achieved via post processing, so that it doesn’t have negative ripple effects throughout the cycled DA.
- Figure 13 was updated with the new results and corresponding discussion was provided.
Figure 14: Please state what the units and physical height correspond to model level 50.
- The model level 50 is a hybrid level which corresponds with around 850 hPa at approximately 1450 m of altitude. As suggested, line 543 was modified for clarity as well as “Hybrid” was included in Figure 14.
Line 543: “...model hybrid level 50 (located in the lower atmosphere at around 850 hPa)”
L 559: “Supersaturation clipping in GSI can improve specific humidity fields in the analyses, allowing for more realistic storm and precipitation forecasts at longer forecast lengths. At shorter forecast lead hours, it produces more spurious convection”
This sounds like the wrong modification is being made to correct a longer-term bias. It would be beneficial to track down the root cause that accumulates to cause the longer-term bias without degrading the short-term skill.
- We agree, however this also highlights a challenge with the partial cycling procedure where we cannot fully examine the model bias owing to the fact that the atmospheric state is periodically refreshed with the global models. Future work involves adapting the approach employed by Wong et al. (2020).
Lines 556-560: “Results from CLIPSAT indicate the presence of longer-term bias that is being corrected to some extent in this experiment. However, because the atmospheric state is periodically refreshed with the large scale conditions as part of the partial cycling procedure, the model bias cannot be fully examined. Further investigation involves adapting the approach employed by Wong et al. (2020) in which forecast tendencies are used to investigate systematic model biases in a continuously cycled experiment.”
L 572: “More extensive testing of RRFS, covering a wider variety of cases, larger domain, and longer period of time, is needed to demonstrate whether results found here are robust or may be case dependent.”
I agree, which is why I’m confused by the presentation of the confidence intervals. Could the authors either remove these in the plots above or describe in greater detail how they were computed and why they are relevant in this case.
- As mentioned above, we agree with the reviewer that the confidence intervals interpretation was not included in the discussion and therefore it was improved in the new version of the manuscript. It is important to highlight that the results presented are representative for the case study analyzed. However, we are not confident that the results obtained in this study can be extrapolated to all other relevant storm instances. Therefore we conclude that more tests covering a wider variety of cases is needed in order to obtain more robust results.
References added to the manuscript in function of this comments and suggestions:
Azevedo, H. B. D., Gonçalves, L. G. G. D., Kalnay, E., and Wespetal, M.: Dynamically weighted hybrid gain data assimilation: perfect model testing, Tellus A: Dynamic Meteorology and Oceanography, 72, 1–11, https://doi.org/10.1080/16000870.2020.1835310, 2020.
Davis, C. A., Brown, B. G., Bullock, R., and Halley-Gotway, J.: The Method for Object-Based Diagnostic Evaluation (MODE)Applied to Numerical Forecasts from the 2005 NSSL/SPC Spring Program, Weather and Forecasting, 24, 1252–1267,https://doi.org/10.1175/2009WAF2222241.1, 2009.
Derber, J. and Rosati, A.: A global oceanic data assimilation system, Journal of physical oceanography, 19, 1333–1347, 1989.
Schwartz, C. S., Poterjoy, J., Carley, J. R., Dowell, D. C., Romine, G. S., and Ide, K.: Comparing Partial and Continuously Cycling Ensemble Kalman Filter Data Assimilation Systems for Convection-Allowing Ensemble Forecast Initialization, Weather and Forecasting, 37, 85 – 112, https://doi.org/10.1175/WAF-D-21-0069.1, 2022
Morris, M. T., Carley, J. R., Colón, E., Gibbs, A., Pondeca, M. S. F. V. D., and Levine, S.: A Quality Assessment of the Real-Time MesoscaleAnalysis (RTMA) for Aviation, Weather and Forecasting, 35, 977 – 996, https://doi.org/10.1175/WAF-D-19-0201.1, 2020.
Wong, M., G. Romine, and C. Snyder, 2020: Model improvement via systematic investigation of physics tendencies. Monthly Weather Review, 148, 671-688. https://doi.org/10.1175/MWR-D-19-0255.1
Zhang, J., Howard, K., Langston, C., Kaney, B., Qi, Y., Tang, L., Grams, H., Wang, Y., Cocks, S., Martinaitis, S., Arthur, A., Cooper, K.,Brogden, J., and Kitzmiller, D.: Multi-Radar Multi-Sensor (MRMS) Quantitative Precipitation Estimation: Initial Operating Capabilities,Bulletin of the American Meteorological Society, 97(4), 621–638, https://doi.org/10.1175/BAMS-D-14-00174.1, 2016.
Citation: https://doi.org/10.5194/gmd-2021-289-AC3
-
AC3: 'Reply on RC3', Ivette Hernandez Banos, 11 Feb 2022