AttentionFire_v1.0: interpretable machine learning fire model for burned area predictions over tropics
- 1Climate and Ecosystem Sciences Division, Climate Sciences Department, Lawrence Berkeley National Laboratory, Berkeley, CA, USA
- 2State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan, China
- 3Department of Civil and Environmental Engineering, University of Illinois Urbana Champaign, Champaign, IL, USA
- 4Department of Earth System Science, University of California Irvine, Irvine, CA, USA
- 5Department of Forest and Wildlife Ecology, University of Wisconsin-Madison, Madison, WI, USA
- 6School of Remote Sensing and Information Engineering, Wuhan University, Wuhan, China
- 1Climate and Ecosystem Sciences Division, Climate Sciences Department, Lawrence Berkeley National Laboratory, Berkeley, CA, USA
- 2State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan, China
- 3Department of Civil and Environmental Engineering, University of Illinois Urbana Champaign, Champaign, IL, USA
- 4Department of Earth System Science, University of California Irvine, Irvine, CA, USA
- 5Department of Forest and Wildlife Ecology, University of Wisconsin-Madison, Madison, WI, USA
- 6School of Remote Sensing and Information Engineering, Wuhan University, Wuhan, China
Abstract. African and South American (ASA) wildfires account for more than 70 % of global burned areas and have strong connection to local climate for sub-seasonal to seasonal wildfire dynamics. However, representation of the wildfire-climate relationship remains challenging, due to spatiotemporally heterogenous responses of wildfires to climate variability and human influences. Here, we developed an interpretable Machine Learning (ML) fire model (AttentionFire_v1.0) to resolve the complex spatial- heterogenous and time-lagged controls from climate on burned area and to better predict burned areas over ASA regions. Our ML fire model substantially improved predictability of burned area for both spatial and temporal dynamics compared with five commonly used machine learning models. More importantly, the model revealed strong time-lagged control from climate wetness on the burned areas. The model also predicted that under a high emission future climate scenario, the recently observed declines in burned area will reverse in South America in the near future due to climate changes. Our study provides reliable and interpretable fire model and highlights the importance of lagged wildfire-climate relationships in historical and future predictions.
-
Notice on discussion status
The requested preprint has a corresponding peer-reviewed final revised paper. You are encouraged to refer to the final revised version.
-
Preprint
(1891 KB)
-
Supplement
(1041 KB)
-
The requested preprint has a corresponding peer-reviewed final revised paper. You are encouraged to refer to the final revised version.
- Preprint
(1891 KB) -
Supplement
(1041 KB) - BibTeX
- EndNote
Journal article(s) based on this preprint
Fa Li et al.
Interactive discussion
Status: closed
-
RC1: 'Comment on gmd-2022-195', Anonymous Referee #1, 25 Aug 2022
General comments:
The authors of “AttentionFire_v1.0: interpretable machine learning fire model for burned area prediction over tropics” developed a novel machine learning model of fire for use in South America and Africa. They use the model to gain insight into the controls of wildfire on the landscape and make future predations. Overall, this paper is interesting and provides insight into an important process and region of the world. I have some major concerns, however, mostly related to the presentation of the methods that should be addressed to improve the clarity and rigor of the manuscript.
1. Adding more explanation about the model and how it related to other machine learning models in plain terms.
2. Providing more information about the data sets used to conduct this analysis to allow the reader to better understand and assess what was done.
3. Make it more clear how the model was validated and include a test against independent data withheld from any tuning to guard against overfitting.
4. Add more explanation about how the future projections were conducted, what input data sets were used, and if and how they were stepped into the future.
5. Provide methods, background, and discussion for section 3.3 which are missing.
6. Address via analysis or discuss the impact of bias in CESM versus the reanalysis data, the impact of coupling between fire, climate, and biomass, and model/scenario uncertainty on the future projections presented.
Specific comments:
L24-27: I suggest also mentioning anthropogenic drivers as they’re included in the mode. Currently, only climate is highlighted.
L41: Placing this emissions number in the context of the carbon budget of these regions using published values could better highlight the importance of this work.
L105-108: I suggest adding more detail here, especially for a reader who is not familiar with machine learning models. That is define black boxed and explain why more complex machine learning models are often less interpretable in straightforward terms. The acronym LSTM should be defined here as well.
L120—130/140-179/figure 1: More background is needed in this section, especially for a reader who is less familiar with artificial neural networks. That is I suggest stating that an LSTM is a type of ANN and explaining its practical advantages and disadvantages versus a typical ANN, and the Naïve LSTM in straightforward terms. Maybe this could also take the form of a table. Figure 1 could be better tied into the text with definitions given for more specific terminology used.
L192-193/198-207: The description of the data sets, their time step (i.e. daily, monthly, etc), units, and origin should be included here. I also suggest moving table S2 into the text and editing it to include more information.
L193-195: Please clarify the validation method used here. Is this a leave-one-out cross-validation? Was any model tuning conducted and what data was used to do that? Ideally, the models would be validated against independent data that was withheld from any tuning/testing to guard against overfitting.
L208-215: These future data need to be prefaced and explained a bit more. Was AttentionFire coupled with CESM or are these simply outputs from CESM? How were variables like road network density and livestock projected into the future? If AttentionFire was not coupled in CESM was bias correction applied to deal with any biases present in the model run, but not the reanalysis? Could these biases impact the results or trends predicted by AttentionFire?
Fig 2: Suggest editing the caption to provide more information about each panel of the figure.
Figure 3: It’s unclear from the figure caption what each of these three panels shows as no letter codes are provided.
Figure 3: Was an attempt made to simplify the model by removing low-ranked data sets? This could be beneficial if it eliminates unimportant variables which are uncertain or hard to obtain in the future
Section 3.3: These experiments are not included in the methods, no background for this is included in the introduction and, acronyms are not defined. Substantial background needs to be added here.
Section 3.4: Several points regarding the future projections are not addressed here. First is the possibility that there is bias in CESM which is not present in the reanalysis data. Should a bias correction be applied? Second, fire, climate, and biomass on the landscape are all coupled. Therefore there is a need to address how this could impact the estimates and trends given if the fire model is not coupled with CESM. Finally, if the models are not coupled only a single model run using a strong emissions scenario is presented here. I’d suggest either presenting additional scenarios and including other models or explaining how model and scenario uncertainty could impact the results, their applicability to this region, and the significant trends highlighted.
Minor comments:
L69: Suggest replacing “from climate” with “of climate”
L70: Suggest replacing “up to multiple” with “on the order of”
L81: Suggest replacing “opposing fire” with “opposite fire”
L212: Suggest adding “the” between “2016-2055” and “99th”
- AC1: 'Reply on RC1', Qing Zhu, 20 Dec 2022
-
RC2: 'Comment on gmd-2022-195', Anonymous Referee #2, 22 Nov 2022
Li et al. presents in this work an attention-augmented LSTM machine learning model framework used to predict burned area over tropics. Attention-augmented models aim to provide interpretability to LSTM models and improve driver selection by adaptively assigning weights to inputs. The authors use this capability to explicitly capture controlling factors of fire predictions with various time-lags (e.g., climate wetness).
I would also like to appreciate the authors’ preparation of the source code to include data preparation scripts and a brief tutorial python script to get users started with the model with examples.
The manuscript is generally well written and the AttentionFire model could be useful for burned area predictions. However, there are some sections which need substantial revisions for clarity and for more details. I will be happy to further consider this manuscript for publication after my concerns are addressed.
Major comments:
1. The authors compare against other models (in Section 2.2), i.e., RF, DT, GBDT, ANN, and naïve LSTM. While Table S1 discusses the hyperparameter configuration of these models, it would be more helpful for model users to read here about the specific strengths and shortcomings of the models chosen – e.g., have these been used for burned area predictions before? Why were these particular models chosen for comparison? Not all models here lack interpretability (DT, RF, …), do they give the same important features as AttentionFire (shown in Fig. 3)? How much more computational cost (memory/data, training time) is incurred with training this more complex, attention-augmented LSTM model, compared to others?Overall, the comparison needs to have more context (for readers who are interested in fire models but not necessarily well-versed in machine learning), and more detail (justifying that the model presented is better and its potential shortcomings). A table similar to Table S1 with a summary of all the models would be helpful in the main text.
2. Section 3.3 regarding oceanic dynamics introduces four oceanic indexes into the model, but they are not defined (except in the legend of Fig. 5) nor introduced. There needs to be more context in this set of experiments: what are these oceanic indexes, why are they important or potentially affecting burned area, how much weight was assigned to these OIs by the model – how much relative importance are they compared to the other predictors, and source of data.
3. Section 3.4 mentions “considering land use changes, population growth, and projected climate under the SSP585”. Is AttentionFire using outputs from a fully coupled CESM simulation? More details about the simulation setup (compsets, any customization to the namelists, input data, etc.) need to be provided (supplement). I understand that the choice for 2016-2055 was due to the model being trained under the historically available data; but I would also suggest testing the AttentionFire model under a different SSP for more complete projections. As it stands section 3.4 heavily leans on the SSP585 CESM model output data for predictions, and the prediction results must be presented carefully, especially when some results are not statistically significant.
4. Finally, the manuscript focuses on predicting ASA wildfires using AttentionFire. Can the AttentionFire model be readily applied for interpretable burned area predictions to other regions with wildfires as well?
Specific comments and technical corrections:
1. Line 110: define LSTM acronym here as its used directly in below text.
2. Line 206: “T62 resolution: 94x192”. I suggest noting the approximate resolution (at equator) here, ~210km, for the spectral resolution.
3. Fig. 2: Are these observations from GFED? It is briefly mentioned in the introduction, but I suggest indicating such within the section 3.1.
4. Fig. 3: Please label the three regions in the figure.
5. Section 3.3: Please avoid defining acronyms in the figure 5 legend, bring them into the main text of section 3.3.
6. Line 426: “will dampened” -> “will be dampened”- AC2: 'Reply on RC2', Qing Zhu, 20 Dec 2022
Peer review completion






Interactive discussion
Status: closed
-
RC1: 'Comment on gmd-2022-195', Anonymous Referee #1, 25 Aug 2022
General comments:
The authors of “AttentionFire_v1.0: interpretable machine learning fire model for burned area prediction over tropics” developed a novel machine learning model of fire for use in South America and Africa. They use the model to gain insight into the controls of wildfire on the landscape and make future predations. Overall, this paper is interesting and provides insight into an important process and region of the world. I have some major concerns, however, mostly related to the presentation of the methods that should be addressed to improve the clarity and rigor of the manuscript.
1. Adding more explanation about the model and how it related to other machine learning models in plain terms.
2. Providing more information about the data sets used to conduct this analysis to allow the reader to better understand and assess what was done.
3. Make it more clear how the model was validated and include a test against independent data withheld from any tuning to guard against overfitting.
4. Add more explanation about how the future projections were conducted, what input data sets were used, and if and how they were stepped into the future.
5. Provide methods, background, and discussion for section 3.3 which are missing.
6. Address via analysis or discuss the impact of bias in CESM versus the reanalysis data, the impact of coupling between fire, climate, and biomass, and model/scenario uncertainty on the future projections presented.
Specific comments:
L24-27: I suggest also mentioning anthropogenic drivers as they’re included in the mode. Currently, only climate is highlighted.
L41: Placing this emissions number in the context of the carbon budget of these regions using published values could better highlight the importance of this work.
L105-108: I suggest adding more detail here, especially for a reader who is not familiar with machine learning models. That is define black boxed and explain why more complex machine learning models are often less interpretable in straightforward terms. The acronym LSTM should be defined here as well.
L120—130/140-179/figure 1: More background is needed in this section, especially for a reader who is less familiar with artificial neural networks. That is I suggest stating that an LSTM is a type of ANN and explaining its practical advantages and disadvantages versus a typical ANN, and the Naïve LSTM in straightforward terms. Maybe this could also take the form of a table. Figure 1 could be better tied into the text with definitions given for more specific terminology used.
L192-193/198-207: The description of the data sets, their time step (i.e. daily, monthly, etc), units, and origin should be included here. I also suggest moving table S2 into the text and editing it to include more information.
L193-195: Please clarify the validation method used here. Is this a leave-one-out cross-validation? Was any model tuning conducted and what data was used to do that? Ideally, the models would be validated against independent data that was withheld from any tuning/testing to guard against overfitting.
L208-215: These future data need to be prefaced and explained a bit more. Was AttentionFire coupled with CESM or are these simply outputs from CESM? How were variables like road network density and livestock projected into the future? If AttentionFire was not coupled in CESM was bias correction applied to deal with any biases present in the model run, but not the reanalysis? Could these biases impact the results or trends predicted by AttentionFire?
Fig 2: Suggest editing the caption to provide more information about each panel of the figure.
Figure 3: It’s unclear from the figure caption what each of these three panels shows as no letter codes are provided.
Figure 3: Was an attempt made to simplify the model by removing low-ranked data sets? This could be beneficial if it eliminates unimportant variables which are uncertain or hard to obtain in the future
Section 3.3: These experiments are not included in the methods, no background for this is included in the introduction and, acronyms are not defined. Substantial background needs to be added here.
Section 3.4: Several points regarding the future projections are not addressed here. First is the possibility that there is bias in CESM which is not present in the reanalysis data. Should a bias correction be applied? Second, fire, climate, and biomass on the landscape are all coupled. Therefore there is a need to address how this could impact the estimates and trends given if the fire model is not coupled with CESM. Finally, if the models are not coupled only a single model run using a strong emissions scenario is presented here. I’d suggest either presenting additional scenarios and including other models or explaining how model and scenario uncertainty could impact the results, their applicability to this region, and the significant trends highlighted.
Minor comments:
L69: Suggest replacing “from climate” with “of climate”
L70: Suggest replacing “up to multiple” with “on the order of”
L81: Suggest replacing “opposing fire” with “opposite fire”
L212: Suggest adding “the” between “2016-2055” and “99th”
- AC1: 'Reply on RC1', Qing Zhu, 20 Dec 2022
-
RC2: 'Comment on gmd-2022-195', Anonymous Referee #2, 22 Nov 2022
Li et al. presents in this work an attention-augmented LSTM machine learning model framework used to predict burned area over tropics. Attention-augmented models aim to provide interpretability to LSTM models and improve driver selection by adaptively assigning weights to inputs. The authors use this capability to explicitly capture controlling factors of fire predictions with various time-lags (e.g., climate wetness).
I would also like to appreciate the authors’ preparation of the source code to include data preparation scripts and a brief tutorial python script to get users started with the model with examples.
The manuscript is generally well written and the AttentionFire model could be useful for burned area predictions. However, there are some sections which need substantial revisions for clarity and for more details. I will be happy to further consider this manuscript for publication after my concerns are addressed.
Major comments:
1. The authors compare against other models (in Section 2.2), i.e., RF, DT, GBDT, ANN, and naïve LSTM. While Table S1 discusses the hyperparameter configuration of these models, it would be more helpful for model users to read here about the specific strengths and shortcomings of the models chosen – e.g., have these been used for burned area predictions before? Why were these particular models chosen for comparison? Not all models here lack interpretability (DT, RF, …), do they give the same important features as AttentionFire (shown in Fig. 3)? How much more computational cost (memory/data, training time) is incurred with training this more complex, attention-augmented LSTM model, compared to others?Overall, the comparison needs to have more context (for readers who are interested in fire models but not necessarily well-versed in machine learning), and more detail (justifying that the model presented is better and its potential shortcomings). A table similar to Table S1 with a summary of all the models would be helpful in the main text.
2. Section 3.3 regarding oceanic dynamics introduces four oceanic indexes into the model, but they are not defined (except in the legend of Fig. 5) nor introduced. There needs to be more context in this set of experiments: what are these oceanic indexes, why are they important or potentially affecting burned area, how much weight was assigned to these OIs by the model – how much relative importance are they compared to the other predictors, and source of data.
3. Section 3.4 mentions “considering land use changes, population growth, and projected climate under the SSP585”. Is AttentionFire using outputs from a fully coupled CESM simulation? More details about the simulation setup (compsets, any customization to the namelists, input data, etc.) need to be provided (supplement). I understand that the choice for 2016-2055 was due to the model being trained under the historically available data; but I would also suggest testing the AttentionFire model under a different SSP for more complete projections. As it stands section 3.4 heavily leans on the SSP585 CESM model output data for predictions, and the prediction results must be presented carefully, especially when some results are not statistically significant.
4. Finally, the manuscript focuses on predicting ASA wildfires using AttentionFire. Can the AttentionFire model be readily applied for interpretable burned area predictions to other regions with wildfires as well?
Specific comments and technical corrections:
1. Line 110: define LSTM acronym here as its used directly in below text.
2. Line 206: “T62 resolution: 94x192”. I suggest noting the approximate resolution (at equator) here, ~210km, for the spectral resolution.
3. Fig. 2: Are these observations from GFED? It is briefly mentioned in the introduction, but I suggest indicating such within the section 3.1.
4. Fig. 3: Please label the three regions in the figure.
5. Section 3.3: Please avoid defining acronyms in the figure 5 legend, bring them into the main text of section 3.3.
6. Line 426: “will dampened” -> “will be dampened”- AC2: 'Reply on RC2', Qing Zhu, 20 Dec 2022
Peer review completion






Journal article(s) based on this preprint
Fa Li et al.
Fa Li et al.
Viewed
HTML | XML | Total | Supplement | BibTeX | EndNote | |
---|---|---|---|---|---|---|
619 | 155 | 12 | 786 | 43 | 7 | 3 |
- HTML: 619
- PDF: 155
- XML: 12
- Total: 786
- Supplement: 43
- BibTeX: 7
- EndNote: 3
Viewed (geographical distribution)
Country | # | Views | % |
---|
Total: | 0 |
HTML: | 0 |
PDF: | 0 |
XML: | 0 |
- 1
The requested preprint has a corresponding peer-reviewed final revised paper. You are encouraged to refer to the final revised version.
- Preprint
(1891 KB) - Metadata XML
-
Supplement
(1041 KB) - BibTeX
- EndNote