Articles | Volume 19, issue 1
https://doi.org/10.5194/gmd-19-27-2026
© Author(s) 2026. This work is distributed under the Creative Commons Attribution 4.0 License.
Increasing resolution and accuracy in sub-seasonal forecasting through 3D U-Net: the western US
Download
- Final revised paper (published on 05 Jan 2026)
- Supplement to the final revised paper
- Preprint (discussion started on 26 Mar 2025)
- Supplement to the preprint
Interactive discussion
Status: closed
Comment types: AC – author | RC – referee | CC – community | EC – editor | CEC – chief editor
| : Report abuse
-
RC1: 'Comment on egusphere-2025-308', Anonymous Referee #1, 25 Jun 2025
- AC1: 'Reply on RC1', Jin-Ho Yoon, 17 Aug 2025
-
RC2: 'Comment on egusphere-2025-308', Anonymous Referee #2, 24 Jul 2025
- AC2: 'Reply on RC2', Jin-Ho Yoon, 17 Aug 2025
Peer review completion
AR – Author's response | RR – Referee report | ED – Editor decision | EF – Editorial file upload
AR by Jin-Ho Yoon on behalf of the Authors (17 Aug 2025)
Author's response
Author's tracked changes
Manuscript
ED: Referee Nomination & Report Request started (28 Aug 2025) by Dan Lu
RR by Anonymous Referee #2 (15 Sep 2025)
ED: Publish subject to technical corrections (16 Nov 2025) by Dan Lu
AR by Jin-Ho Yoon on behalf of the Authors (22 Nov 2025)
Manuscript
Summary
The paper proposes and tests a method for downscaling sub-seasonal weather forecasts to improve accuracy and spatial resolution. The approach uses a neural network architecture (3D U-Net) that has previously been used for similar tasks. The forecasts are from the ECWMF ensemble forecast system and the high-resolution data are from PRISM. The method is applied in the Western US. The effect of different input data (ensemble members versus mean, different variable sets) are examined. The authors find that the neural network improves temperature predictions relative to the original NWP forecasts. Results for precipitation are also presented but the quality is more mixed.
High level feedback
The topic itself is interesting, and the results shown (particularly for temperature) seem promising. However, it is difficult to evaluate the method due to the omission of important details. Some of the missing methodological details can be deduced by reading the code, but they should be provided in the paper itself.
In additional to technical omissions, a main concern is the lack of clarity around the purpose and impact of the analysis. The main question, as I understand it, is: what is the effect of using (a) different sets of predictor variables and (b) different ensemble components/aggregations on prediction accuracy? The introduction states that Höhlein et al. (2024) examined (b) and reached approximately the same conclusion as this study. How is this study different? Question (a) seems relevant. However, there is very little discussion of the specific predictor variables (I think they are only mentioned in the SI) and how/why specific variables may contribute to better or worse predictions. Again, it is not clearly explained how this study differs from the cited Horat and Lerch (2024) or Weyn et al. (2021). Most of the framing of the results and conclusions boil down to “neural network downscaling improves prediction accuracy relative to NWP,” which, based on the introduction, seems to already be well-established.
The motivation for the ensemble-based predictors is also confusing. The purpose of an ensemble prediction system is to represent uncertainty, which is not discussed. Also the ensemble members are simulations that, by construction, do not start from the “optimal” estimate of the initial conditions. So it is not surprising that E01 performs worse (unless by “first” ensemble member you mean the control). Interpreting the relative performance of E50 versus E50M requires methodological details that are not provided. But again it is not surprising that the performance is similar given that E50 output is being reduced to a deterministic prediction. It seems like the value of downscaling based on an ensemble would be more in representing forecast uncertainty than improving deterministic downscaled predictions.
Specific feedback
Minor comments
Line 16 (and elsewhere): when you refer to a paper in running text, you should still provide the year in parentheses
Line 34: Re “subset of variables,” a subset of what? I think you’re referring to “additional” or maybe “auxiliary” variables, but I don’t think subset is the right term. If anything it’s a (super)set that includes the target variables as a subset.
Line 36: Again I do not think “sub-variable” is the right term here. See above for suggested alternatives.
Lines 49-51: On line 49 is says “we select forecasts from CY40R1” and on line 51 it says “We utilize forecasts from CY40R1 to CY48R1.” This might be clearer to someone more familiar with ECMWF forecast naming scheme, but I find this confusing. It would be helpful to give some additional explanation what this means and maybe provide a link to the relevant data product(s).
Line 70: Remove “properly.” Also, I believe the preferred GMD style is “Fig.” rather than “Figure” in running text.
Line 73: I generally find it clearer to talk in terms of fine versus coarse spatial resolution.
Line 96: Regarding “conservative interpolation,” can you be more specific about the method?
Lines 97-98: Regarding “given the established relationship …”, what is the relationship?
Line 110: semicolon should be colon
First paragraph of 3.1: This description of the scope of analysis should come much earlier in the paper, not in the results section.
Line 131: I think you mean “metrics”
Line 141: “suggest elevation can enhance temperature post-processing accuracy” this is just lapse rate, right? I don’t think this should be framed as a finding of NN methods
References
Entekhabi, D., Reichle, R.H., Koster, R.D. and Crow, W.T., 2010. Performance metrics for soil moisture retrievals and application requirements. Journal of Hydrometeorology, 11(3), pp.832-840.