the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
CRITER 1.0: A coarse reconstruction with iterative refinement network for sparse spatio-temporal satellite data
Abstract. Satellite observations of sea surface temperature (SST) are essential for accurate weather forecasting and climate modeling. However, this data often suffers from incomplete coverage due to cloud obstruction and limited satellite swath width, which requires development of dense reconstruction algorithms. The current state-of-the-art struggles to accurately recover high-frequency variability, particularly in SST gradients in ocean fronts, eddies, and filaments, which are crucial for downstream processing and predictive tasks. To address this challenge, we propose a novel two-stage method CRITER (Coarse Reconstruction with ITerative Refinement Network), which consists of two stages. First, it reconstructs low-frequency SST components utilizing a Vision Transformer-based model, leveraging global spatio-temporal correlations in the available observations. Second, a UNet type of network iteratively refines the estimate by recovering high-frequency details. Extensive analysis on datasets from the Mediterranean, Adriatic, and Atlantic seas demonstrates CRITER's superior performance over the current state-of-the-art. Specifically, CRITER achieves up to 44 % lower reconstruction errors of the missing values and over 80 % lower reconstruction errors of the observed values compared to the state-of-the-art.
- Preprint
(9906 KB) - Metadata XML
- BibTeX
- EndNote
Status: open (extended)
-
CEC1: 'Comment on gmd-2024-208', Juan Antonio Añel, 21 Mar 2025
reply
Dear authors,
Unfortunately, after checking your manuscript, it has come to our attention that it does not comply with our "Code and Data Policy".
https://www.geoscientific-model-development.net/policies/code_and_data_policy.htmlYour code availability section points out to a Zenodo repository that is empty, and does not contain the CRITER 1.0 code.
Therefore, the current situation with your manuscript is irregular. Please, publish your code in one of the appropriate repositories and reply to this comment with the relevant information (link and a permanent identifier for it (e.g. DOI)) as soon as possible, as we can not accept manuscripts in Discussions that do not comply with our policy.
In this way, if you do not fix this problem, we will have to reject your manuscript for publication in our journal.
Also, you must include a modified 'Code and Data Availability' section in a potentially reviewed manuscript, containing the DOI of the new repositories. Also, neither on the Zenodo repository nor on GitHub a license is listed. If you do not include a license, the code is not "free-libre open-source (FLOSS)"; it remains your property. Therefore, when uploading the model's code to Zenodo, you could want to choose a free software/open-source (FLOSS) license. You could want to use the GPLv3. You simply need to include the file 'https://www.gnu.org/licenses/gpl-3.0.txt' as LICENSE.txt with your code. Also, you can choose other options that Zenodo provides: GPLv2, Apache License, MIT License, etc.
Juan A. Añel
Geosci. Model Dev. Executive EditorCitation: https://doi.org/10.5194/gmd-2024-208-CEC1 -
AC1: 'Reply on CEC1', Matjaž Zupančič Muc, 22 Mar 2025
reply
Dear executive editor,
Thank you for your feedback. We apologize for the oversight in our code availability section.
We have published the CRITER 1.0 code on Zenodo under the MIT license. It is now available at: https://doi.org/10.5281/zenodo.13923156
We will update the manuscript's 'Code and Data Availability' section accordingly.
Thank you for your consideration.
Sincerely,
Matjaž Zupančič MucCitation: https://doi.org/10.5194/gmd-2024-208-AC1 -
CEC2: 'Reply on AC1', Juan Antonio Añel, 22 Mar 2025
reply
Dear authors,
Many thanks for addressing this issue so quickly. We can consider now the current version of your manuscript in compliance with the policy of the journal.
Juan A. Añel
Geosci. Model Dev. Executive Editor
Citation: https://doi.org/10.5194/gmd-2024-208-CEC2
-
CEC2: 'Reply on AC1', Juan Antonio Añel, 22 Mar 2025
reply
-
AC1: 'Reply on CEC1', Matjaž Zupančič Muc, 22 Mar 2025
reply
-
RC1: 'Comment on gmd-2024-208', Anonymous Referee #1, 01 May 2025
reply
Review of "CRITER 1.0: A coarse reconstruction with iterative refinement network for sparse spatio-temporal satellite data"
by Matjaž Zupancic Muc, Vitjan Zavrtanik, Alexander Barth, Aida Alvera-Azcarate, Matjaž Licer, and Matej KristanThis manuscript is a description of a novel machine learning technique for gap filling of SST analysis in the presence of possibly significant missing (satellite) observational data. The reconstruction method uses high resolution, multi-sensor, binned where observations exist, L3S SST products from Copernicus Marine Service, and then uses a two stage approach machine learning technique to both fill the true missing data as well as further missing data removed from the L3S product to be used for training and validation. The analysis is then validated against this removed data, showing improvements over other methods -- primarily DINEOF of Alvera-Azcárate et al., (2005).
Firstly, I am not an expert on machine learning techniques, and therefore will offer limited comment of the techniques involved, but rather a potential user of improved SST analysis, and therefore offer comments more aligned with that perspective. This is one of my major points of commentary on the present manuscript: The paper as a whole is a rather technical description of the proposed method -- and rightly so. However, I believe some additional commentary on potential users of the system, and what benefits it might offer them should be addressed in the introduction. As it stands now, this is only addressed very briefly and casually in literally the first 4 lines of the introduction, after which the manuscript pivots to solely detailing the technical details. My additional major comment would be to better describe some of the terminology in the manuscript. The meaning of seemingly simple terminology, such as that used for variance, as well as deleted and visible regions, is likely inherently obvious to the authors, however, the interpretation of these terms by the reader could lead to some confusion. Some more detailed descriptions with regards to the used terminology may be necessary, even if this seems painfully obvious to the authors.
Major Comments:
1) Not enough motivating background information in the introduction. Other than the first 4 lines of the introduction, no motivating information is provided as to why improved high-resolution SST reconstructions are necessary. While everyone would presumably like the best possible SST reconstruction, what applications would best benefit, and how might they benefit? Although more directed towards satellite capabilities than gap filling techniques, a review article such as "Observational Needs of Sea Surface Temperature" (https://doi.org/10.3389/fmars.2019.00420) would seem a good starting point for building motivation. Other articles exploring the use of improving the resolution of SST boundary conditions for numerical weather prediction could also prove useful. A quick search yielded me these two possibilities (10.1175/JCLI-3275.1, 10.5194/hess-24-269-2020). Presumably a more detailed background search would yield more.
- Given that high resolution global NWP systems -- ECMWF's IFS is 9km (1/12o) -- better high resolution global SST products are also required. The SST reconstructions pursued in this manuscript are all regional (Mediterranean, Adriatic, North Atlantic). It is not mentioned whether it would be practical to scale the proposed technique to global domains, such as gap filling the Copernicus Marine Service 1/10o ODYSSEY L3 product.
2) Some terminology used in the manuscript, while seemingly obvious, on further contemplation the meaning and interpretation is not so obvious.
- Uncertainty/Variance (σ2): The uncertainty or variance outcome from the machine learning training process is introduced and summarized with the generic statement leading off section 3 in the opening 3 lines (ll. 82-84). This statement represents the only description of how this quantity, which plays a large role in the analysis of the techniques performance and skill over the rest of the manuscript. If possible a more detailed description of how this term is output or diagnosed from the machine learning process would be warranted. From a naive aspect, I would assume this variance, or uncertainty is the range of SST values that would lead to the same best fit outcome in the training process, but obviously, not enough information is given to confirm this. Furthermore, as detailed in the paper on "Observational Needs of Sea Surface Temperature" given above, and the outcome of many workshops on the needs required of SST observations and analysis, there is a strong need for estimates of uncertainty to accompany estimates of SST. The estimate of variance/uncertainty outcome from this technique seems well posed to fulfill this requirement -- if its definition is an adequate measure of this.
- The definition of variance becomes further confused with the introduction of scaled error (error divided by variance, l. 251, 3rd line of S5.3). While the authors again use symbol σ for the scaled variance, or more precisely, σε, this is well identified. The confusion (for me) was then when scaled variance , σε << 1 was compared with an idealized reconstruction where σε = 1, this is casually referred to as an overestimate of the variance (ll. 261-262). It took me more than a few moments to eventually realize this was the scaled variance, with the actual variance being a divisor to this scaled variance -- and therefore scaled variance , σε < 1, does indeed represent an overestimation of actual variance. At the risk of insulting some all knowing readers, but lifting up some of the slower to comprehend readers, please somehow remind the readers that this is the scaled variance which is divided by the actual variance -- and therefore the statement does actually makes sense.
- Visible and Deleted regions: The definition of deleted regions seem relatively obvious: The regions where SST observations have artificially been removed from the L3 product. However, the definition of visible, sometimes referred to as observed, regions seems less definite: Is it the fully observed region in the L3 SST before deletion, or the observed region in the L3 SST after removal of the deleted regions? Please provide a precise definition of deleted and visible regions.
Typographic and style comments:
Section 4 Results (l. 168) is empty?
Figures 4-6, B1-B3: Limits on σ and rmse. The colour scale limits on σ and rmse seem to be all automatically generated. This is a hindrance to both comparing between techniques (CRITER/DINCAE2) and comparing over-dispersive and under-dispersive regions (σ vs rmse). Although I realize this will often lead to regions of colour saturation, I would strive (at least for individual scenarios) to have the colour scale range identical between CRITER/DINCAE2 results and between σ and rmse (preferably with the zero value always represented). This would likely enhance your ability to discuss the results in the text, and by setting the scales for σ and rmse identically, it would then allow you to connect the results in Section 5.3 with the earlier results -- for instance, you would easily be able to identify regions where DINCAE2 has insufficient variance compared to RMSE, and vice versa for CRITER).
l. 263 : in order of 10−2◦ or lower. Not sure what is meant by to the power of -2o? Possible typographical error. "In order of" is more conventionally referred to as "Of the order of"
Further Comments:
Comments outside scope of manuscript that may be worthy of at least some discussion.
1) As already mentioned, I do not know much about technical details of machine learning techniques. But what I do know is that the techniques are relatively agnostic to the physical relationship between the inputs and outputs. In this study, one inputs binned temperature observations and outputs the full field temperatures. The input L3S products used in the study have undergone a variety of processing (radiance algorithms, bias corrections) to produce a binned multi-sensor temperature. Would this technique be generalizable to the underlying radiances, complicated by requiring different trainings for each instrument? The advantage might, however, be better instrument bias corrections and a further reduction in error?
2) As mentioned in Major Comment #1: Is this scalable to a global analysis?
3) It could be interesting to apply a (spatial) spectral analysis on the results and underlying inputs, which admittedly would likely require large cloud free areas, at least for analyzing the spectral characteristics of the inputs. Do the wavelength characteristics of the CRITER and DINCAE2 results differ, and how do they compare to the original wavelength characteristics of the binned SST L3S products: Are certain wavelengths removed and/or enhanced.
Citation: https://doi.org/10.5194/gmd-2024-208-RC1
Viewed
HTML | XML | Total | BibTeX | EndNote | |
---|---|---|---|---|---|
200 | 42 | 11 | 253 | 13 | 13 |
- HTML: 200
- PDF: 42
- XML: 11
- Total: 253
- BibTeX: 13
- EndNote: 13
Viewed (geographical distribution)
Country | # | Views | % |
---|
Total: | 0 |
HTML: | 0 |
PDF: | 0 |
XML: | 0 |
- 1