the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
SICNetseason V1.0: a transformer-based deep learning model for seasonal Arctic sea ice prediction by integrating sea ice thickness data
Abstract. The Arctic sea ice is suffering dramatic retreating in summer and fall, which has far-reaching consequences on the global climate and commercial activities. Accurate seasonal sea ice predictions are significant in inferring climate change and planning commercial activities. However, seasonally predicting the summer sea ice encounters a significant obstacle known as the spring predictability barrier (SPB): predictions made later than May demonstrate good skill in predicting summer sea ice, while predictions made on or earlier than May exhibit considerably lower skill. This study develops a transformer-based deep-learning model, SICNetseason (V1.0), to predict the Arctic sea ice concentration on a seasonal scale. Including spring sea ice thickness (SIT) data in the model significantly improves the prediction skill at the SPB point. A 20-year (2000–2019) testing demonstrates that the detrended anomaly correlation coefficient (ACC) of Sep. sea ice extent (sea ice concentration > 15 %) predicted by our model at May/Apr. is improved by 7.7 %/10.61 % over the ACC predicted by the state-of-the-art dynamic model from the European Centre for Medium-Range Weather Forecasts (ECMWF). Compared with the anomaly persistence benchmark, the mentioned improvement is 41.02 %/36.33 %. Our deep learning model significantly reduces prediction errors of Sep.'s sea ice concentration on seasonal scales compared to ECMWF and Persistence. The spring SIT data is key in optimizing the SPB, contributing to a more than 20 % ACC enhancement in Sep.'s SIE at four to five months lead predictions. Our model achieves good generalization in predicting the Sep. SIE of 2020–2023.
- Preprint
(4732 KB) - Metadata XML
- BibTeX
- EndNote
Status: open (until 15 Feb 2025)
-
CEC1: 'Comment on gmd-2024-200', Juan Antonio Añel, 27 Dec 2024
reply
Dear authors,
Unfortunately, after checking your manuscript, it has come to our attention some lack of compliance with our "Code and Data Policy".
https://www.geoscientific-model-development.net/policies/code_and_data_policy.htmlTo assure the replicability of your work it is necessary to have all the input and output data that you use and produce, respectively. You have provided a repository hosted by Zenodo for part of the input data and the code. However, in the repository it reads "Because the well-trained model is too large (10GB), it cannot be supported by the github". I understand that this is a sentence that comes from a previously existing GitHub site where you hosted these assets. However, it is clear that does not correspond to the current Zenodo repository. Moreover, Zenodo does not have problems with hosting 10 GB of data, so you must include in it all the input data that you have used. Also, in your "Code and Data Availability" section you link several webpages that provide access to the data that you use for your work. These are not valid and should be replaced by the repositories containing the exact input data. The sites that you list are only main "portals" to data, not the repositories for the exact input data that you use, and what you have to provide is the exact input data used to produce your work. Moreover, such sites are not valid repositories for long-term archival or scientific publication.
Additionally, you do not provide the output data produced by your model, and you must do it.
Therefore, you must reply to this comment as soon as possible with the relevant information (link and DOI) for the new repositories, as we request that you make it already available before submission and, of course, before the Discussions stage. Moreover, you must include in a potentially revised version of your manuscript the modified 'Code and Data Availability' section, with the DOI of the code (and another DOI for the dataset if necessary).
Note that if you do not fix these problems as requested, we will have to reject your manuscript for publication in our journal. In the meantime, I advice the Topical Editor to stall the peer-review process until these issues are solved, as this manuscript should not have been accepted in Discussion given the lack of compliance with the policy of the journal.
Juan A. Añel
Geosci. Model Dev. Executive Editor
Citation: https://doi.org/10.5194/gmd-2024-200-CEC1 -
AC1: 'Reply on CEC1', Yibin Ren, 27 Dec 2024
reply
Thanks for the comment.
We have uploaded the exact input data, output data, and the well-trained model weights to the new Zenodo repository with a DOI: https://doi.org/10.5281/zenodo.14561423. We have added detail introductions for the upload files to make sure the model can be run:
- The codes of the model are in the model_utils folders.
- The jupyter file "SICNet_season_training.ipynb" contains the flow of loading data, defining a model, and training the model.
- The jupyter file "SICNet_season_evaluating.ipynb" includes evaluating the trained model and drawing the ACC/BACC result.
- The training and testing data are in the Data/Data.zip file and can be loaded for training and evaluation of the model. The files "SIC_1979_2020_Jan_May_96.npy", "SIT_1979_2020_Jan_May_96.npy", "SIC_climate_mean_96.npy", and "mask_96.npy" are input data for model training. The "Output data" file contains the predicted results output by SICNetseason model, "Predicted_SIC_output.npy". The "Ground_SIC_output.npy" in this file is the ground truth from NSIDC.
- The Model_SICNet.zip contains the saved well-trained models. Please unpackage the zip file and copy the .h5 files to the corresponding file path "/Result/Model_SICNet". Then, the "SICNet_season_evaluating.ipynb" can be run to genernate the predicted values by the well-trained models.
We also revised the “Code and data availability” session as follow:
The satellite-observed sea ice concentration is obtained from the following site https://nsidc.org/data/NSIDC-0081/versions/1 (Cavalieri et al., 1996). The reanalysis of sea ice thickness data is open access (http://psc.apl.uw.edu/research/projects/arctic-sea-ice-volume-anomaly/data/model_grid) (Zhang and Rothrock, 2003). The seasonal predictions of SEAS5 are obtained from the ECMWF (https://cds.climate.copernicus.eu/cdsapp#!/dataset/seasonal-monthly-single-levels?tab=form) (Johnson et al., 2019). The code, the exact input/output data, and the saved well-trained weights of the developed model SICNetseason are available at https://doi.org/10.5281/zenodo.14561423.
We upload the revised manuscript with the new “Code and data availability” section as supplement.
-
CEC2: 'Reply on AC1', Juan Antonio Añel, 28 Dec 2024
reply
Dear authors,
Thanks for replying so quickly to my previous comment. However, I do not understand why you continue including in the Code and Data Availability section links to the sites that do not comply with the policy. It is my understanding that all the input data is in the Zenodo repository, and therefore, linking other sites is unnecessary and only generates doubts among readers about where to find the data to replicate your work. Therefore, the section should read simply "The code, the exact input/output data, and the saved well-trained weights of the developed model SICNetseason are available at https://doi.org/10.5281/zenodo.14561423."
If my interpretation is wrong, and it is necessary to use data that is not included in the Zenodo repository, and that is hosted in the other sites, then your manuscript is not in compliance with the policy of the journal, because such sites are not accepted as suitable repositories for scientific publication.
Please, clarify the situation regarding the input data, and delete the mentions to the sites that do not comply with the policy.
Juan A. Añel
Geosci. Model Dev. Executive Editor
Citation: https://doi.org/10.5194/gmd-2024-200-CEC2 -
AC2: 'Reply on CEC2', Yibin Ren, 28 Dec 2024
reply
Thanks for the comment. Sorry for my misunderstanding of the last comment. We have deleted other sites and revised the "Code and data availability" section as:
The code, the exact input/output data, and the saved well-trained weights of the developed model SICNetseason are available at https://doi.org/10.5281/zenodo.14561423.
The new revised manuscript is attached with the supplement.
-
CEC3: 'Reply on AC2', Juan Antonio Añel, 28 Dec 2024
reply
Dear authors,
Many thanks for the quick reply again. We can consider now the current version of your manuscript in compliance with our code and data policy, and therefore the Discussions stage and the peer-review process can continue.
Juan A. Añel
Geosci. Model Dev. Executive Editor
Citation: https://doi.org/10.5194/gmd-2024-200-CEC3
-
CEC3: 'Reply on AC2', Juan Antonio Añel, 28 Dec 2024
reply
-
AC2: 'Reply on CEC2', Yibin Ren, 28 Dec 2024
reply
-
AC1: 'Reply on CEC1', Yibin Ren, 27 Dec 2024
reply
-
RC1: 'Comment on gmd-2024-200', Anonymous Referee #1, 09 Jan 2025
reply
For comments - see pdf attached
-
AC3: 'Reply on RC1', Yibin Ren, 22 Jan 2025
reply
General comments:In summary, this research aims to develop a transformer-based U-net model to improve predictionskill of Arctic sea ice concentration at seasonal lead times. The primary novel contribution of thiswork is the use of a state-of-the-art deep learning architecture with a custom domain-specific lossfunction as well as including spring sea ice thickness as an input predictor to overcome the commonspring predictability barrier (SPB) faced by previous numerical and DL-based models that decreaseprediction skill of predictions made before May.The authors choose to use a transformer based architecture to capture spatiotemporal patterns in theinput data. The model inputs include multiple sea-ice related variables over 3-6 months and outputsthe sea ice concentration of the next six months, specifically looking at predictions for June-Sept at6 month lead times. The authors develop a novel loss function that combines both a standard DLloss function (mean squared error) and a domain-specific NIEE loss which accounts for spatialsimilarity in model predictions and ground truth. The model follows a standard DL trainingprocedure and includes three years of unseen data to test their model performance. The authorsevaluate their model using standard DL performance metrics along with Binary Accuracy (BACC)which accounts for accurate spatial distributions.The authors succeed in improving prediction skill when compared to the benchmark Persistencemodel and dynamical ECMWF model. For Sept ice prediction specifically, the ECMWF modelperforms better at smaller lead times but the author’s model shows marked performanceimprovement at a lead time of 3-6 months, thus overcoming the spring predictability barrier. Theauthors perform necessary ablation tests to investigate the role of sea ice thickness in their model.They show a model that includes sea ice thickness as input outperforms a model without sea icethickness at seasonal lead times of 3-6 months. When testing their model on unseen data, the authorsuse an ensemble model approach by averaging the results of 20 separately trained models. Theauthors also test their model against a previous state of the art deep learning CNN based approach,IceNet. They show their model produces better ACC scores at longer lead times. Their modelproduced lower BACC scores but can capture more individualized/local or extreme characteristicsin comparison to IceNet which tends to produce smoothed out results.Overall, this manuscript provides a significant improvement to the modeling of sea ice concentrationby providing a novel approach to increase prediction skill at seasonal lead times to overcome theSPB. The authors clearly state their motivations for the project and outline their novel contributions.The authors discuss previous modeling approaches in the field and clearly showcase their significantscientific results where their model outperforms previous approaches. The contributions of thispaper is twofold, one in utilizing a novel transformer based approach that the authors state has notbeen previously used in this field and second in highlighting the impact of spring sea ice thicknesson improving prediction accuracies, revealing potential scientific insights that need to be studiedfurther. Below are a few suggestions to improve reproducibility and presentation quality.Response: Thanks for the comment.Specific comments:Comment 1: In line 52, the authors state that experiments show the reason for decrease inprediction skill before spring is due to ice motion and growth in the winter. It is unclear whichexperiments in the literature the authors are referring to. Including a citation here would helpjustify this statement.Response: Thanks for the comment. We cite the following reference in this sentence:Bushuk, M., Winton, M., Bonan, D. B., Blanchard-Wrigglesworth, E., and Delworth, T. L.: Amechanism for the Arctic sea ice spring predictability barrier, Geophys Res Lett, 47,https://doi.org/10.1029/2020GL088335, 2020.Comment 2: In line 79, a citation for the IceNet model (Andersson et al 2021) is missing. Sinceworking with the IceNet model is a significant portion of the manuscript, perhaps includingmore information about the IceNet model and how it differentiates from the author’stransformer based model would provide more context to the reader in this section.Response: Thanks for the comment. We added the citation for IceNet in line 79. A brief introductionabout IceNet has also been added. The revised sentences are as follows:Finally, we compare our SICNetseason model with the published deep learning model IceNet(Andersson et al., 2021). IceNet is a probability prediction model for Arctic SIE based onconvolutional neural network (CNN) units and the U-Net architecture. It achieved state-of-the-artperformance in predicting the SIE for six months (Andersson et al., 2021). Therefore, we choseIceNet as a comparison model.Besides, more information about the IceNet model, such as the model’s inputs and outputs, waspresented in Section 4.7, lines 263-267:The IceNet is a seasonal sea ice prediction model that performs state-of-the-art SIE prediction(Andersson et al., 2021). It is a CNN-based classification model, and it outputs the probability ofthree classes: open water (SIC≤15%), marginal ice (15% < SIC < 80%), and full ice (SIC≥80%).Differently, our SICNetseason outputs the 0-100% range SIC values. The IceNet's inputs consist of 50monthly mean variables, including SIC, 11 climate variables, statistical SIC forecasts, and metadata.Comment 3: In lines 108-109, the authors state that multiple experiments were conducted todetermine the length of the model input features. It is unclear what experiments or methods theauthors tried. Was this determined by manual tuning using their cross-validation strategy or did theauthors employ an automated grid-search type strategy. The authors can also include if domainknowledge or previous literature informed the choice of input lengths of the data.Response: Thanks for the comment. We determine the length of input factors by combining domainknowledge and manual tuning experiments. The main domain knowledge we considered is the seaice reemergence mechanism. The spring-fall reemergence occurs between pairs of months wherethe ice edge is in the same position, such as in May and December (Blanchard-Wrigglesworth et al.,2011; Day et al., 2014). The spring sea ice anomaly is positively correlated with fall sea iceanomalies, and there is also a weaker reemergence between fall sea ice anomalies and anomaliesthe following spring (Bushuk et al., 2015). Therefore, we set the initial input length of theSIC/SIT/SIC anomaly as six months. Then, we change the length manually to fine-tune the deep learning model to find the best-matched length for each factor. The SIC climatology of the targetmonths provides an essential mean state of the prediction SIC, so we input it into the model. It canalso represent the month number signal that IceNet has considered. We explain more details aboutthis issue in the revised manuscript, and the new revision is as follows:The input for SICNetseason is a 96×96×18 SIC and SIT sequence, composed of SIT of the last threemonths, SIC of the last six months, SIC anomaly of the last three months, and SIC climatology ofthe six target months (Fig. 1a). We determine the length of input factors by combining domainknowledge and manual tuning experiments. The primary domain knowledge we considered is thespring-fall reemergence mechanism. It occurs between pairs of months where the ice edge is in thesame position, such as in May and December (Blanchard-Wrigglesworth et al., 2011; Day et al.,2014). The spring sea ice anomaly is positively correlated with fall sea ice anomalies, and there isalso a weaker reemergence between fall sea ice anomalies and anomalies the following spring(Bushuk et al., 2015). Therefore, we set the initial input length of the SIC/SIT/SIC anomaly as sixmonths. We change the input length manually (from six to one in step one) to fine-tune the deeplearning model to find the best-matched length for each factor. The SIC climatology of the targetmonths provides an essential mean state of the prediction SIC. It represents the monthly cycle signalthat IceNet has considered.[1] Blanchard-Wrigglesworth, E., Armour, K. C., Bitz, C. M., and Deweaver, E.: Persistence andinherent predictability of arctic sea ice in a GCM ensemble and observations, J Clim, 24, 231–250,https://doi.org/10.1175/2010JCLI3775.1, 2011.[2] Day, J. J., Tietsche, S., and Hawkins, E.: Pan-arctic and regional sea ice predictability:Initialization month dependence, J Clim, 27, 4371–4390, https://doi.org/10.1175/JCLI-D-13-00614.1, 2014.[3] Bushuk, M., Giannakis, D., and Majda, A. J.: Arctic Sea Ice Reemergence: The Role of LargeScale Oceanic and Atmospheric Variability*, https://doi.org/10.1175/JCLI-D-14-00354.s1, 2015.Comment 4: In section 4.7, when comparing the transformer based model to the CNN-basedIceNet model, the authors state they used identical training and testing settings to perform faircomparisons. It is unclear whether the authors used the same 20 trained ensemble modelapproach they had used for their transformer model for the IceNet model. If so, specifyingwhether they used an ensemble approach or singular-model approach for IceNet would clarifythis for the reader. If the authors did not use a similar ensemble approach, the authors shouldjustify this choice.Response: Thanks for the comment. We are sorry for the confusion. We did not use an ensembleapproach for both SICNetseason and IceNet. The training procedure is a leave-one-year-out strategyfor the 20 testing years (2000-2019). For example, if the testing year is 2019, the training set is datafrom 1979-2018, and the testing data is 2019. Then, the testing year moves to 2018, and thecorresponding training set is data from 1979-2017 and 2019. The model is trained 3 times for eachtraining-testing pair to eliminate randomness, and the prediction for each testing year is the meanvalue of the three trained models. We explain the leave-one-year-out training procedure in Section4.1. Further, we clarify the training strategy in the revision:The training and testing settings of IceNet are the same as those of SICNetseason. The IceNet is trainedusing the same leave-one-year-out strategy as the SICNetseason. For example, if the testing year is 2019, the training set is data from 1979-2018, and the testing data is 2019. Then, the testing datamoves to 2018; the remaining data (1979-2017, 2019) is the training set. For each training/testingpair, the model is trained three times to eliminate randomness, and the final prediction for testingdata is the mean value of the three models.Comment 5: In line 269, it is unclear how the authors transformed the IceNet output to matchthe continuous scale (for e.g restructuring only the final output layer). Having this additionalcontext would help with reproducibility of this experiment.Response: Thanks for the comment. We reconstruct IceNet's output layer by replacing the softmaxwith the sigmoid activation function. The sigmoid function outputs continuous values of 0-100%.We clarify this point in the revision:We reconstruct IceNet’s output layer by replacing the original softmax with the sigmoid activationfunction. The sigmoid function outputs continuous values of 0-100%.Comment 6: For Figures 2, 4, 6 and 8, it is slightly confusing to the reader at first how tointerpret these results, specifically the Difference column. Perhaps including the caption thatred signifies improvement in accuracy and blue signifies a decrease would aid in understandingespecially because in Figures 3 and 5 the opposite color scheme is used (red = high error, blue– lower error).Response: Thanks for the comment. We have added the following statement to the captions ofFigures 2, 4, 6, and 8: The red signifies a high/improvement in ACC/BACC, and the blue signifiesa decrease.
-
AC3: 'Reply on RC1', Yibin Ren, 22 Jan 2025
reply
Viewed
HTML | XML | Total | BibTeX | EndNote | |
---|---|---|---|---|---|
211 | 45 | 10 | 266 | 1 | 2 |
- HTML: 211
- PDF: 45
- XML: 10
- Total: 266
- BibTeX: 1
- EndNote: 2
Viewed (geographical distribution)
Country | # | Views | % |
---|
Total: | 0 |
HTML: | 0 |
PDF: | 0 |
XML: | 0 |
- 1