Search

Machine-learning-based evidence and attribution mapping of 100000 climate impact studies - Nature.com

jokbanga.blogspot.com

Abstract

Increasing evidence suggests that climate change impacts are already observed around the world. Global environmental assessments face challenges to appraise the growing literature. Here we use the language model BERT to identify and classify studies on observed climate impacts, producing a comprehensive machine-learning-assisted evidence map. We estimate that 102,160 (64,958–164,274) publications document a broad range of observed impacts. By combining our spatially resolved database with grid-cell-level human-attributable changes in temperature and precipitation, we infer that attributable anthropogenic impacts may be occurring across 80% of the world’s land area, where 85% of the population reside. Our results reveal a substantial ‘attribution gap’ as robust levels of evidence for potentially attributable impacts are twice as prevalent in high-income than in low-income countries. While gaps remain on confidently attributabing climate impacts at the regional and sectoral level, this database illustrates the potential current impact of anthropogenic climate change across the globe.

Access options

Rent or Buy article

Get time limited or full article access on ReadCube.

from$8.99

All prices are NET prices.

Fig. 1: Results of the machine-assisted literature review.
Fig. 2: Potential attribution of impact studies to regional anthropogenic temperature and precipitation trends.
Fig. 3: A global density map of climate impact evidence.

Data availability

The results of this study are made available in a public repository58.

Code availability

The code used to produce these results is made available in a public repository50.

References

  1. 1.

    Cramer, W. et al. in Climate Change 2014: Impacts, Adaptation, and Vulnerability (eds Field, C. B. et al.) 979–1037 (Cambridge Univ. Press, 2014).

  2. 2.

    IPCC Climate Change 2014: Impacts, Adaptation, and Vulnerability (eds Field, C. B. et al.) (Cambridge Univ. Press, 2014).

  3. 3.

    Hansen, G. The evolution of the evidence base for observed impacts of climate change. Curr. Opin. Environ. Sustain. 14, 187–197 (2015).

    Article  Google Scholar 

  4. 4.

    Haunschild, R., Bornmann, L. & Marx, W. Climate change research in view of bibliometrics. PLoS ONE 11, e0160393 (2016).

    Article  Google Scholar 

  5. 5.

    Grieneisen, M. L. & Zhang, M. The current status of climate change research. Nat. Clim. Change 1, 72–73 (2011).

    Article  Google Scholar 

  6. 6.

    Haddaway, N. R. & Pullin, A. S. The policy role of systematic reviews: past, present and future. Springer Sci. Rev. 2, 179–183 (2014).

    Article  Google Scholar 

  7. 7.

    Callaghan, M. W., Minx, J. C. & Forster, P. M. A topography of climate change research. Nat. Clim. Change 10, 118–123 (2020).

    Article  Google Scholar 

  8. 8.

    Porciello, J., Ivanina, M., Islam, M., Einarson, S. & Hirsh, H. Accelerating evidence-informed decision-making for the Sustainable Development Goals using machine learning. Nat. Mach. Intell. 2, 559–565 (2020).

    Article  Google Scholar 

  9. 9.

    Nunez-Mir, G. C., Iannone, B. V. III, Curtis, K. & Fei, S. Evaluating the evolution of forest restoration research in a changing world: a “big literature” review. New For. 46, 669–682 (2015).

    Article  Google Scholar 

  10. 10.

    Westgate, M. J. et al. Software support for environmental evidence synthesis. Nat. Ecol. Evol. 2, 588–590 (2018).

    Article  Google Scholar 

  11. 11.

    Lamb, W. F., Creutzig, F., Callaghan, M. W. & Minx, J. C. Learning about urban climate solutions from case studies. Nat. Clim. Change 9, 279–287 (2019).

    Article  Google Scholar 

  12. 12.

    Cohen, A. M. An effective general purpose approach for automated biomedical document classification. AMIA Annu. Symp. Proc. 2006, 161–165 (2006).

    Google Scholar 

  13. 13.

    Marshall, I. J., Kuiper, J., Banner, E. & Wallace, B. C. Automating biomedical evidence synthesis: RobotReviewer. In Proc. Association for Computational Linguistics Meeting 7–12 (The Association for Computational Linguistics, 2017).

  14. 14.

    Baclic, O. et al. Challenges and opportunities for public health made possible by advances in natural language processing. Can. Commun. Dis. Rep. 46, 161–168 (2020).

    Article  Google Scholar 

  15. 15.

    Schleussner, C.-F. & Fyson, C. L. Scenarios science needed in UNFCCC periodic review. Nat. Clim. Change 10, 272 (2020).

    Article  Google Scholar 

  16. 16.

    Fankhauser, S. Adaptation to climate change. Annu. Rev. Resour. Econ. 9, 209–230 (2017).

    Article  Google Scholar 

  17. 17.

    Bedsworth, L. W. & Hanak, E. Adaptation to climate change. J. Am. Plann. Assoc. 76, 477–495 (2010).

    Article  Google Scholar 

  18. 18.

    IPCC Special Report on Managing the Risks of Extreme Events and Disasters to Advance Climate Change Adaptation (Cambridge Univ. Press, 2012).

  19. 19.

    Hallegatte, S. & Mach, K. J. Make climate-change assessments more relevant. Nature 534, 613–615 (2016).

    CAS  Article  Google Scholar 

  20. 20.

    Conway, D. et al. The need for bottom-up assessments of climate risks and adaptation in climate-sensitive regions. Nat. Clim. Change 9, 503–511 (2019).

    Article  Google Scholar 

  21. 21.

    Hansen, G. & Stone, D. Assessing the observed impact of anthropogenic climate change. Nat. Clim. Change 6, 532–537 (2016).

    Article  Google Scholar 

  22. 22.

    Knutson, T. R., Zeng, F. & Wittenberg, A. T. Multimodel assessment of regional surface temperature trends: CMIP3 and CMIP5 twentieth-century simulations. J. Clim. 26, 8709–8743 (2013).

    Article  Google Scholar 

  23. 23.

    Knutson, T. R. & Zeng, F. Model assessment of observed precipitation trends over land regions: detectable human influences and possible low bias in model trends. J. Clim. 31, 4617–4637 (2018).

    Article  Google Scholar 

  24. 24.

    Nerem, R. S. et al. Climate-change-driven accelerated sea-level rise detected in the altimeter era. Proc. Natl Acad. Sci. USA 115, 2022–2025 (2018).

    CAS  Article  Google Scholar 

  25. 25.

    Gudmundsson, L., Leonard, M., Do, H. X., Westra, S. & Seneviratne, S. I. Observed trends in global indicators of mean and extreme streamflow. Geophys. Res. Lett. 46, 756–766 (2019).

    Article  Google Scholar 

  26. 26.

    PadrĂłn, R. S. et al. Observed changes in dry-season water availability attributed to human-induced climate change. Nat. Geosci. 13, 477–481 (2020).

    Article  Google Scholar 

  27. 27.

    Devlin, J., Chang, M.-W., Lee, K. & Toutanova, K. BERT: pre-training of deep bidirectional transformers for language understanding. Preprint at https://arxiv.org/abs/1810.04805 (2019).

  28. 28.

    Sanh, V., Debut, L., Chaumond, J. & Wolf, T. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. Preprint at https://arxiv.org/abs/1910.01108 (2020).

  29. 29.

    Halterman, A. Mordecai: full text geoparsing and event geocoding. J. Open Source Softw. 2, 91 (2017).

    Article  Google Scholar 

  30. 30.

    Lane, J. E., Kruuk, L. E. B., Charmantier, A., Murie, J. O. & Dobson, F. S. Delayed phenology and reduced fitness associated with climate change in a wild hibernator. Nature 489, 554–557 (2012).

    CAS  Article  Google Scholar 

  31. 31.

    Zhang, Y. Q., Yu, C. H. & Bao, J. Z. Acute effect of daily mean temperature on ischemic heart disease mortality: a multivariable meta-analysis from 12 counties across Hubei Province, China. Zhonghua Yu Fang Yi Xue Za Zhi 50, 990–995 (2016).

    CAS  Google Scholar 

  32. 32.

    Barry, A. A. et al. West Africa climate extremes and climate change indices. Int. J. Climatol. 38, e921–e938 (2018).

    Article  Google Scholar 

  33. 33.

    Hegerl, G. C. et al. Good practice guidance paper on detection and attribution related to anthropogenic climate change. In Meeting Report of the Intergovernmental Panel on Climate Change Expert Meeting on Detection and Attribution of Anthropogenic Climate Change (eds Stocker, T. F. et al.) (IPCC, 2010).

  34. 34.

    Rosenzweig, C. et al. in Climate Change 2007: Impacts, Adaptation and Vulnerability (eds Parry, M. L. et al.) 79–131 (Cambridge Univ. Press, 2007).

  35. 35.

    Rosenzweig, C. et al. Attributing physical and biological impacts to anthropogenic climate change. Nature 453, 353–357 (2008).

    CAS  Article  Google Scholar 

  36. 36.

    Gridded Population of the World, Version 4 (GPWv4): Population Density Revision 11 (CIESIN, 2018).

  37. 37.

    Frank, D. et al. Effects of climate extremes on the terrestrial carbon cycle: concepts, processes and potential future impacts. Glob. Change Biol. 21, 2861–2880 (2015).

    Article  Google Scholar 

  38. 38.

    Schleussner, C.-F. et al. 1.5 °C hotspots: climate hazards, vulnerabilities, and impacts. Annu. Rev. Environ. Resour. 43, 135–163 (2018).

    Article  Google Scholar 

  39. 39.

    Peng, R. D. Reproducible research in computational science. Science 334, 1226–1227 (2011).

    CAS  Article  Google Scholar 

  40. 40.

    MĂĽller-Hansen, F., Callaghan, M. W. & Minx, J. C. Text as big data: develop codes of practice for rigorous computational text analysis in energy social science. Energy Res. Soc. Sci. 70, 101691 (2020).

    Article  Google Scholar 

  41. 41.

    Shepherd, T. G. Storyline approach to the construction of regional climate change information. Proc. R. Soc. A 475, 20190013 (2019).

    Article  Google Scholar 

  42. 42.

    Rosenzweig, C. & Neofotis, P. Detection and attribution of anthropogenic climate change impacts. Wiley Interdiscip. Rev. Clim. Change 4, 121–150 (2013).

    Article  Google Scholar 

  43. 43.

    Mengel, M., Treu, S., Lange, S. & Frieler, K. ATTRICI 1.1—counterfactual climate for impact attribution. Geosci. Model Dev. https://doi.org/10.5194/gmd-14-5269-2021 (2021).

  44. 44.

    Gudmundsson, L. et al. Globally observed trends in mean and extreme river flow attributed to climate change. Science 371, 1159–1162 (2021).

    CAS  Article  Google Scholar 

  45. 45.

    Diffenbaugh, N. S. Verification of extreme event attribution: using out-of-sample observations to assess changes in probabilities of unprecedented events. Sci. Adv. 6, eaay2368 (2020).

    Article  Google Scholar 

  46. 46.

    Herring, S. C., Christidis, N., Hoell, A., Hoerling, M. P. & Stott, P. A. Explaining Extreme Events of 2019 from a Climate Perspective (American Meteorological Society, 2021).

  47. 47.

    Cochrane Handbook for Systematic Reviews of Interventions (John Wiley & Sons, 2019).

  48. 48.

    Callaghan, M., MĂĽller-Hansen, F., Hilaire, J. & Lee, Y. T. NACSOS: NLP assisted classification, synthesis and online screening. Zenodo https://doi.org/10.5281/zenodo.4121526 (2020).

  49. 49.

    McHugh, M. L. Interrater reliability: the kappa statistic. Biochem. Med. 22, 276–282 (2012).

    Article  Google Scholar 

  50. 50.

    Callaghan, M. Machine learning-based evidence and attribution mapping of 100,000 climate impact studies - code. Zenodo https://doi.org/10.5281/ZENODO.5327409 (2021).

  51. 51.

    Chang, C.-C. & Lin, C.-J. LIBSVM: a library for support vector machines. ACM Trans. Intell. Syst. Technol. 2, 27 (2011).

    Article  Google Scholar 

  52. 52.

    Bender, E. M., Gebru, T., McMillan-Major, A. & Shmitchell, S. On the dangers of stochastic parrots: can language models be too big? In Proc. ACM Conference on Fairness, Accountability, and Transparency 610–623 (Association for Computing Machinery, 2021); https://doi.org/10.1145/3442188.3445922

  53. 53.

    Bergstra, J. & Bengio, Y. Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13, 281–305 (2012).

    Google Scholar 

  54. 54.

    Gururangan, S. et al. Don’t stop pretraining: adapt language models to domains and tasks. Preprint at https://arxiv.org/abs/2004.10964 (2020).

  55. 55.

    Morice, C. P., Kennedy, J. J., Rayner, N. A. & Jones, P. D. Quantifying uncertainties in global and regional temperature change using an ensemble of observational estimates: the HadCRUT4 data set. J. Geophys. Atmos. https://doi.org/10.1029/2011JD017187 (2012).

  56. 56.

    Eyring, V. et al. Overview of the Coupled Model Intercomparison Project Phase 6 (CMIP6) experimental design and organization. Geosci. Model Dev. 9, 1937–1958 (2016).

    Article  Google Scholar 

  57. 57.

    Beusch, L., Gudmundsson, L. & Seneviratne, S. I. Crossbreeding CMIP6 Earth system models with an emulator for regionally optimized land temperature projections. Geophys. Res. Lett. 47, e2019GL086812 (2020).

    Article  Google Scholar 

  58. 58.

    Callaghan, M. et al. Machine learning-based evidence and attribution mapping of 100,000 climate impact studies - data. Zenodo https://doi.org/10.5281/ZENODO.5257271 (2021).

Download references

Acknowledgements

M.C. is supported by a PhD stipend from the Heinrich Böll Stiftung. J.C.M. acknowledges funding from the ERC-2020-SyG GENIE (grant ID 951542). S.N. and Q.L. acknowledge funding from the German Federal Ministry of Education and Research (BMBF) and the German Aerospace Center (DLR) via the LAMACLIMA project as part of AXIS, an ERANET initiated by JPI Climate (http://www.jpi-climate.eu/AXIS/Activities/LAMACLIMA, last access: 26 August 2021, grant no. 01LS1905A), with co-funding from the European Union (grant no. 776608). M.R. acknowledges support by the ERC-SyG USMILE (grant ID 85518). R.J.B. acknowledges support from the EU Horizon2020 Marie-Curie Fellowship Program H2020-MSCA-IF-2018 (proposal no. 838667 -INTERACTION). We thank F. Zeng for providing preliminary temperature and precipitation trend assessment results for our project. We acknowledge the World Climate Research Programme, which, through its Working Group on Coupled Modelling, coordinated and promoted CMIP6. We thank the climate modelling groups for producing and making available their model output, the Earth System Grid Federation (ESGF) for archiving the data and providing access and the multiple funding agencies who support CMIP6 and ESGF.

Author information

Affiliations

Authors

Contributions

M.C., J.C.M. and C.-F.S. designed the research. M.C. developed the coding platform and machine-learning pipeline to identify studies, with advice from M.R. M.C, C-F.S., G.H., Q.L. and E.T. developed the codebook and coordinated screening and coding. M.C., Q.L., S.N. and C-F.S. conceptualized the link to detection and attribution data. S.N. performed the univariate detection and attribution analysis of temperature and precipitation trends and assessment of internal variability, in consultation with T.R.K, who designed the methodology for these calculations. M.C. and S.N. designed and implemented the matching of studies with detection and attribution data. M.C., C-F.S., S.N., Q.L, G.H., E.T., M.A., R.J.B., M.H., C.J., K.L., A.L., N.v.M., I.M., P.P. and B.Y. contributed to screening and coding studies. M.C., C-F.S., J.C.M., Q.L. and S.N. wrote the manuscript with contributions from all authors.

Corresponding author

Correspondence to Max Callaghan.

Ethics declarations

Competing interests

The authors declare no competing interests.

Additional information

Peer review information Nature Climate Change thanks Abeed Sarker and the other, anonymous, reviewer(s) for their contribution to the peer review of this work.

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Extended data

Extended Data Fig. 1 A visual representation of the workflow of our machine learning assisted attribution map.

Squares represent documents (not to scale), boxes represent the steps taken. Documents are screened by hand, and those labels are used to generate predictions and machine label documents. These machine-labelled documents are matched by location with information from observations and climate models on the detection and attribution of trends in temperature and precipitation.

Extended Data Fig. 2 Nested cross validation (CV) procedure for the binary relevance classifier.

Models are fit using training documents and evaluated on validation/test documents. The inner CV loop is used to search for optimal hyperparameter settings, which are then evaluated on the outer test sets.

Extended Data Fig. 3 Performance metrics for the binary inclusion/exclusion classifier.

Each pair of dots represents the scores for a distinct cross-validation fold. Horizontal lines show the mean score across folds.

Extended Data Fig. 4 Receiver operating curve area under the curve scores (ROC AUC) and F1 scores for the classification of impact categories.

Each pair of dots represents the scores for a distinct cross-validation fold. Horizontal lines show the mean score across folds.

Extended Data Fig. 5 Receiver operating curves area under the curve scores (ROC AUC)(ROC) and F1 scores for the classification of drivers.

Each pair of dots represents the scores for a distinct cross-validation fold. Horizontal lines show the mean score across folds.

Extended Data Fig. 6 Geographical distribution of surface trends.

Temperature from 1951 to 2018 (left) and precipitation trends from 1951 to 2016 (right) in (a),(b) observations and (c),(d) CMIP6 10-model ensemble mean all-forcing runs. Bottom panels (e),(f) show observations categorised into attribution categories, following refs. 8,7, respectively. Observed cooling/warming or drying/wetting trends that–after accounting for internal climate variability–are inconsistent with the simulated response to natural forcings but consistent with the simulated response to both natural and anthropogenic forcings are indicated by categories -/+2. This is clearest case of changes that are at least partially attributable to anthropogenic forcing, according to the CMIP6 ensemble. Categories -/+1 have detectable observed changes, but are not assessed as attributable to anthropogenic forcing because the observed changes are significantly less than those simulated in the average all-forcing runs. Categories -/+3 have detectable changes and are assessed as at least partly attributable anthropogenic forcing, although the observed changes are inconsistent with the all-forcing runs. That is, they are in the same direction as, but are significantly stronger than, the mean of the all-forcing runs. Categories -/+4 represents cooling/warming or drying/wetting trends that are inconsistent with the simulated response to natural forcings but whose sign is opposite to that of the average simulated all-forcing response; category 0 represents trends that are not distinguishable from natural variability alone. Categories -/+4 and 0 are considered to be examples of non-detectable trends).

Extended Data Fig. 7 Fractional difference between average CMIP6 modeled low-frequency standard deviation of annual mean precipitation vs observed precipitation.

To estimate the internal low-frequency variability for both models and observations, the observed time series were detrended and low-pass filtered with a 7-year running mean filter prior to computing the standard deviations while for the models we used the full available control runs (7-yr running mean filtered) to estimate the internal low-frequency variability for each model. The top panel shows the multi-model ensemble standard deviation comparison while the ten individual panels below it show the comparison for each individual CMIP6 model used in the study. The fraction difference was computed as: [(Model st. dev. - Observed st. dev.) / (Observed st. dev.)].

Extended Data Fig. 8 Difference between average CMIP6 modeled low-frequency standard deviation (°C) of annual mean surface air temperature vs observed surface temperature.

To estimate the internal low-frequency variability for both models and observations, the observed time series were detrended and low-pass filtered with a 7-year running mean filter prior to computing the standard deviations while for the models we used the full available control runs (7-year running mean filtered) to estimate the internal low-frequency variability for each model. The top panel shows the multi-model ensemble standard deviation comparison while the ten individual panels below it show the comparison for each individual CMIP6 model used in the study.

Extended Data Fig. 9 An illustration of the spatial resolution and weighting methodology.

Detection and attribution categories for temperature in East Africa; b. the number of grid cells of each type in Sudan; c. weighted studies for each grid cell in Sudan; d. The number of studies referring to each extracted geographical location in Sudan.

Supplementary information

Supplementary Information

Query and Supplementary Fig. 1.

Supplementary Data

A list of the categories used to code relevant documents. Each category could be used as an impact or a driver. To make the classification problem tractable, the categories were merged into ‘broad categories’, resembling those used in IPCC AR5, and ‘aggregated categories’, which distinguish between different types of impacts to a greater extent.

Rights and permissions

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Callaghan, M., Schleussner, CF., Nath, S. et al. Machine-learning-based evidence and attribution mapping of 100,000 climate impact studies. Nat. Clim. Chang. (2021). https://ift.tt/3BsB9vg

Download citation

Adblock test (Why?)



"machine" - Google News
October 11, 2021 at 10:06PM
https://ift.tt/3iSZhQB

Machine-learning-based evidence and attribution mapping of 100000 climate impact studies - Nature.com
"machine" - Google News
https://ift.tt/2VUJ7uS
https://ift.tt/2SvsFPt

Bagikan Berita Ini

0 Response to "Machine-learning-based evidence and attribution mapping of 100000 climate impact studies - Nature.com"

Post a Comment

Powered by Blogger.