Articles | Volume 18, issue 23
https://doi.org/10.5194/gmd-18-9897-2025
https://doi.org/10.5194/gmd-18-9897-2025
Methods for assessment of models
 | 
10 Dec 2025
Methods for assessment of models |  | 10 Dec 2025

Validation of climate mitigation pathways

Pascal Weigmann, Rahel Mandaroux, Fabrice Lécuyer, Anne Merfort, Tabea Dorndorf, Johanna Hoppe, Jarusch Muessel, Robert Pietzcker, Oliver Richters, Lavinia Baumstark, Elmar Kriegler, Nico Bauer, Falk Benke, Chen Chris Gong, and Gunnar Luderer
Abstract

Integrated assessment models (IAMs) are crucial for climate policymaking, offering climate mitigation scenarios and contributing to IPCC assessments. However, IAMs face criticism for lack of transparency and limited ability to represent recent technology diffusion and dynamics. We introduce the Potsdam Integrated Assessment Modeling validation tool, piamValidation, an open-source R package for validating IAM scenarios. The piamValidation tool enables systematic comparisons of variables from extensive IAM datasets against historical data and feasibility bounds or across scenarios and models. This functionality is particularly valuable for harmonizing scenarios across multiple IAMs. Moreover, the tool facilitates the systematic comparison of near-term technology dynamics with external observational data, including historical trends, near-term developments, and empirical findings. We apply the tool in two application cases. First, to scenarios from the Network for Greening the Financial System (NGFS) to demonstrate its general applicability, and second to the integrated assessment model REMIND for near-term technology trend validation, illustrating its potential to enhance the transparency and reliability of IAMs. We apply the tool in two application cases. First, to scenarios from the Network for Greening the Financial System (NGFS) to demonstrate its general applicability, and second to the integrated assessment model REMIND for near-term technology trend validation, illustrating its potential to enhance the transparency and reliability of IAMs.

Share
1 Introduction

Integrated assessment models (IAMs) play a prominent role in providing science-based energy assessments and climate policy advice. Early IAM applications date back to the early 1990s (Cointe et al.2019) and evolved towards the formulation of mitigation targets and the monitoring of political ambition (Van Beek et al.2020). IAM scenarios are also an important pillar of the assessments conducted by the Intergovernmental Panel on Climate Change (IPCC), in particular with regard to transformation pathways towards achieving climate policy goals. Therefore, they are a central component of climate change mitigation-oriented policies. Edenhofer and Minx (2014) argue that the Summary for Policymakers (SPM) established a collaborative environment where political discussions link with relevant scientific material. They conceptualize scientists as mapmakers in the territory of climate policy. In this interpretation, the role of the Conference of the Parties (COP) can be understood through a metaphor, “the COP operates as navigator, navigating a terrain charted by the IPCC” (Beck and Oomen2021, p. 172). In AR6 WGIII Chap. 3, advances in climate modelling are assessed by examining the capability of models to accurately simulate historical developments, underscoring the growing importance of scenario validation in IPCC processes.

Considering the prominent role of IAMs in the climate science-policy interface, IAMs have faced increasing scrutiny. Particularly in policy advice, central criticism ranges from shortcomings regarding model representation (e.g., heterogeneous actors and capital markets) to their ability to capture technology diffusion and dynamics (Keppo et al.2021). On the one hand, global and national models have often underestimated the rapid technological change in renewable power and demand electrification technologies: for instance, the cost of solar electricity and battery storage has declined by almost 90 % (Creutzig et al.2023) and renewable electricity generation in Asia has more than doubled (IRENA2024a) in the last decade, far exceeding model assumptions and results. National models face the additional challenge of making assumptions about technology innovation and cost declines, which are largely driven by global developments and trends. Accurately capturing recent trends in technology capital costs is therefore crucial, as they play a predominant role in the sensitivity of model outputs (Giannousakis et al.2021). On the other hand, near-term IAM scenario results exceeded the rate at which economies are transitioning away from fossil fuels: for example, due to factors beyond pure market-based energy economics, coal power capacities are still being built in some regions despite the declining renewable energy costs. In particular, China and India alone account for 86 % of worldwide coal power capacity under development (Carbon Brief2024).

Not only techno-economic parameters but also the technology representations more generally impact technology trends in IAM scenarios and vary strongly between IAMs, as illustrated by Krey et al. (2019) for the case of electricity generation. They first highlight the need for caution in harmonizing techno-economic assumptions, as they must fit model-specific technology representations and associated projection methods. Second, Krey et al. (2019) participate in the transparency debate (Stanton et al.2009) by urging the IAM community to publish and openly discuss techno-economic parameters, their definitions, and model technology representations.

This paper introduces piamValidation, the Potsdam Integrated Assessment Modeling validation tool for IAMs, to foster its broad adoption in the IAM community and enhance transparency and reliability.

Building on the evaluation framework of Wilson et al. (2021), the piamValidation tool facilitates evaluation of historical simulations, near-term empirical validation, stylized fact checks, and model intercomparison, thereby it can be used to contribute to the appropriateness, interpretability, relevance, and especially the credibility of IAM results. This tool and the related input data are meant to be a community resource, where the distribution of the piamValidation configuration files can serve as a knowledge-sharing platform and as a performance metric, offering continuous feedback on modeling progress.

In Sect. 2, we present the piamValidation open-source R package. The tool enables users to systematically compare variables of large IAM datasets with historical data and feasibility bounds or across scenarios and models. The core function of the tool is to validate IAM results and identify discrepancies. In addition, it also supports harmonizing national and global IAMs in the short and medium term, particularly for current policy scenarios that capture implemented government policies. Furthermore, the tool allows for systematic comparisons of near-term technology dynamics with external observational data as historical, near-term trends or empirical estimates. The piamValidation tool is particularly user-friendly and requires only a single command to generate the full HTML report featuring validation heat maps.

In Sect. 3, we present two complementary application cases of the piamValidation tool. The first, based on scenarios of the Network for Greening the Financial System (NGFS), illustrates its general applicability across diverse contexts, with an emphasis on demonstrating application types rather than results. The second applies piamValidation to evaluate and strengthen the near-term realism of technology trends in the IAM REMIND (REgional Model of Investment and Development, see Baumstark et al. (2021) for details). In the REMIND case study we show how the piamValidation tool can be used to first spot the deviations to historical and outlook data and then demonstrate performance improvements near-term realism focusing on offshore wind capacity, carbon management, and electric cars. The rationale for choosing these technologies is twofold: first, the feasibility of their near-term scaling has severe implications for scenarios with high ambition for emissions reduction (Brutschin et al.2021; Bertram et al.2024); second, the piamValidation tool highlighted relatively large deviations in these variables between the REMIND scenarios and historical data as well as current technology trends. We then present strategies that have been implemented to improve REMIND, and we demonstrate that subsequent scenarios improve their near-term realism.

2 PIAM validation tool

2.1 Overview

The package piamValidation provides validation tools for the Potsdam Integrated Assessment Modeling environment. Developed in R, this open-source package is freely available on GitHub under the LGPL-3 license (https://github.com/pik-piam/piamValidation, last access: 5 December 2025), facilitating transparency and collaboration in research and analysis.

Users provide the tool with IAM scenario data and relevant reference data for historical or future time periods. Criteria for checks are defined in a configuration file to perform a vetting. This configuration file serves as a flexible interface, which allows for various use cases:

  • Deviations of scenario to reference data

  • Deviations between:

    • Models (same scenario, same periods);

    • Scenarios (same model, same periods);

    • Periods (same model, same scenario).

  • Conformity to scalar thresholds, such as an absolute value or growth rate.

Deviations can be defined as either absolute or relative deviations. The function validateScenarios takes the inputs mentioned above and evaluates the validation checks. In the standard setting, this results in a dataset that contains an evaluation of either “green” (test passed), “yellow” (warning), or “red” (fail) for each data point to be evaluated. To further enhance visualization, additional colors can be employed to distinguish between exceedance of lower bounds and upper bounds. In that case, “blue” and “cyan” are used to hint toward values below thresholds, and “yellow” and “red” remain to capture exceeded upper thresholds.

The function validationHeatmap makes these results accessible by arranging them for one variable on an interactive heat map using the R library ggplotly. Depending on the category of checks performed, the underlying data used in the evaluation can be viewed for each data point by hovering over the corresponding square in the heat map. Finally, validationReport can be used to perform the validation calculation and plot all heat maps in one HTML report, which is automatically rendered from an RMarkdown file provided by the package.

After successfully installing the piamValidation R package in an appropriate R environment, the tool structure can be described along three dimensions: Input, data processing, and output, as illustrated in Fig. 1.

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-f01

Figure 1Validation workflow with respective inputs, processing steps and outputs.

Download

2.2 Data Input

The input data for the piamValidation tool comprises three key components. First, the IAM scenario data to be validated. Second, the reference data, which is used to compare the initial IAM scenario data. This reference data may include outputs from other IAM models or scenarios, as well as third-party external datasets, such as historical, observational or other prediction data. Third, a configuration file is necessary to specify the details of the validation. The piamValidation R package provides a standard set of configuration files for general use. However, users seeking greater flexibility can create and customize their own configuration files to address specific validation requirements.

To facilitate use by various IAM communities and external users, IAM data are organized according to the standardized Integrated Assessment Consortium (IAMC) format (https://www.iamconsortium.org/scientific-working-groups/data-protocols-and-management/iamc-time-series-data-template/, last access: 5 December 2025). It organizes data along the dimensions of “model”, “scenario”, “variable”, “region” and “period”. Thus, the validation process works with 5-dimensional data objects, expecting a data point to be uniquely defined as x(m,s,v,r,p), with the indices referring to these five dimensions.

2.2.1 Configuration File

The piamValidation framework allows for a wide range of different validation approaches. They are defined via a configuration file with predefined columns as described in Table 1. This is intended to enable users with limited programming expertise to use the tool, as the structure of the validation configuration file follows the intuitive framework of a validation process, answering the following three guiding questions:

  1. Which type of validation should be performed using which metric?

  2. Which data should be validated?

  3. What is the reference being used for the validation?

These questions are answered by filling in the columns of the configuration file as described in Table 1.

Table 1Column description of configuration file. Note: Each described column must be filled in accordance to the use case to ensure successful operation of the piamValidation tool.

Download Print Version | Download XLSX

2.2.2 Validation types and metrics

A common validation procedure consists of comparing IAM results to benchmarks1, such as observational data from external data sources. Such reference data can be handed to the validation tool in large quantities by supplying a CSV, Excel, or MIF file that is structured along the same five dimensions as the IAM data, with the specific requirement that the “scenario” column reads “historical” for all data points. This way, the tool can also differentiate between a model intercomparison exercise, where the same scenario is compared between two or more models, and a comparison of multiple scenarios to one or multiple reference models. In terms of data dimensions, a relative deviation can thus be expressed as

rel_deviation=x(mi,sj,vk,rl,pm)-x(mref,shist,vk,rl,pm)x(mref,shist,vk,rl,pm),

where x(mi,sj,vk,rl,pm) are the IAM values using the notation introduced in Sect. 2.2 and x(mref,shist,vk,rl,pm) are the observational data values from one or multiple reference models. In an analogous way, a “difference to historical” check can be performed by defining an absolute deviation to the reference data as

abs_deviation=x(mi,sj,vk,rl,pm)x(mref,shist,vk,rl,pm).

Similarly, checks with the metric “relative” or “difference” can also be performed against subsets of the scenario data itself, which then functions as reference data. This can in turn be performed in three different ways – comparing to a selected period, scenario, or model:

abs_deviation_period=x(mi,sj,vk,rl,pm)x(mi,sj,vk,rl,pref),abs_deviation_scen=x(mi,sj,vk,rl,pm)x(mi,sref,vk,rl,pm)orabs_deviation_model=x(mi,sj,vk,rl,pm)x(mref,sj,vk,rl,pm).

Relative comparisons again divide this difference by the reference value. Only one dimension in period, scenario, or model can be used as a validation dimension, while the other two must remain constant. Accordingly, it is not possible to compare scenario sj of model mi with scenario sref of model mref). The reference of this dimension can be either a single period/scenario/model or a comma-separated list of them. Furthermore, when choosing multiple sources the user can decide whether the validation thresholds should be based on the range of the reference values, where the maximum value of the group is used as the reference for upper and the minimum as the reference for lower thresholds or their mean will serve as reference for all thresholds.

For instance, in a near-term validation study, multiple data sources can be chosen as references by selecting the following expression in the config file as ref_model2:

range(mi,mj,,mx).

The metric “absolute” does not require any reference data, as it contains the absolute values of the thresholds directly in the threshold columns of the configuration file.

Lastly, the “growthrate” metric can be used to perform validation checks with respect to average yearly growth rates. For a time step of dt between periods, the growth rate is calculated as

growthrate=x(mi,sj,vk,rl,pm)x(mi,sj,vk,rl,pm-dt)1dt-1.

Depending on the metric of the validation check and whether it is a comparison to observational data, different columns of the configuration file need to be filled as shown in Fig. 2.

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-f02

Figure 2Overview of required and optional columns of the configuration file, depending on metric.

Download

2.3 Data Processing

The validation configuration file allows users to define a wide range of validation checks, going beyond the ones provided by default in the validation tool. On the one hand, this requires the data processing and evaluation to be both consistent and flexible, e.g. when performing data harmonization as n intermediate calculation steps might be required in some validation exercises. On the other hand, users might supply the validation tool with data and configuration files that are either internally inconsistent or exceed the scope of the currently supported use cases. In such cases, it is crucial to provide transparency by ensuring the user can identify the issue through clear mechanisms and the inclusion of essential information in error messages.

These requirements are met by the modular structure of the tool, where each data processing step is carefully separated from the next. The structure of the central data processing function of the validation tool, validateScenarios, is sketched using pseudocode in Algorithm 1.

At the beginning of a validation process using the validateScenarios function, the configuration file and scenario and reference data are imported and checked for consistency (see Algorithm 1). The loop starting in line 4 takes information from one row of the configuration file and assembles the scenario and reference data needed for the validation check defined. It connects the thresholds that are relevant for each data point to the scenario data. In case a validation check is performed against a reference value (either historical or period/scenario/model intercomparison), this value is appended for each data point. The resulting data slices for all rows are combined and checked for any duplicates that can result from overlapping definitions in the configuration file. Lines 11 and 12 contain the calculation and evaluation of the actual validation check, according to the respective metrics. Finally, data can be either exported as CSV or returned as an R data.frame.

Algorithm 1validateScenarios is the core of the piamValidation framework.

Input: scenarioData, referenceData, validationConfig
Output: data.frame or CSV file containing original data, reference values and validation results
function validateScenarios()
1:
data = read(scenarioData, referenceData)
2:
cfg = read(validationConfig)
3:
cfg = check, clean and expand cfg
4:
for all row in cfg do
5:
slice data according to filters in row
6:
left join config thresholds to data slice
7:
left join reference values to data slice (if applicable)
8:
valiData = bind all data slices
9:
end for
10:
valiData = removeDuplicateRows(valiData)
11:
calculate value to be checked against threshold (depending on metric)
12:
evaluate check and append validation result (color) to valiData
13:
(optional: export valiData as CSV file)
14:
return(valiData)

2.4 Output and Visualization

While the data output of validateScenarios can be used in any preferred way, two specific visualization approaches are offered to users of piamValidation. The function validationHeatmap provides a quick overview of the general outcome of a validation by laying out the validation result colors across up to four dimensions. This means that the five-dimensional IAM data objects need to be reduced to a single element in only one dimension. By default, the “variable” dimension is chosen for this. An example can be seen in Fig. 3 where “region” and “period” are chosen for the x- and y-axis respectively and “scenario” and “model” are plotted as x- and y-facets.

Alternatively, the function linePlotsThresholds offers an easy way to plot the thresholds as colored background areas behind line plots of the scenario data and, if applicable, the corresponding reference data. For this type of plot, one additional data dimension needs to be singular, by default, this falls to the “region” dimension.

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-f03

Figure 3Example of an interactive validation heat map. Data along four dimensions. Hovering over a data point shows additional information in a tooltip. In this example, the Final Energy (FE) in two versions of the REMIND model for two different policy scenarios (National Policies implemented (NPi) and “well below” 2 °C (66 % likelihood) of temperature rise by the end of the century (2C) are compared with the historical IEA data.

Download

The heat map plotting routine makes use of the ggplot2 library using the geom_tile function in combination with plotly which enables interactive access to additional information by showing tooltips when hovering over a tile. This allows users to first get a quick overview of the general validation outcome by scanning the colors of the heat map and identifying areas of interest (e.g., a region with many red tiles). In a second step, the exact value of a data point as well as the corresponding thresholds and reference values, including their origin, are shown in the tooltip. This two-step process is a particular strength of the chosen visualization approach, as it makes validation results of large datasets accessible without overcrowding the plot or distributing the data over many different plots (line plots are usually not useful for visualizing data along four dimensions, often it is the ”region” dimension which is spread over multiple plots).

The heat map tiles are chosen to always be squares, which results in the plot layout being dependent on the lengths of the data dimensions. In cases where a high number of scenarios or models is being validated, this can lead to gaps between heat map facets and inefficient use of space in the validation document. To alleviate this, validationHeatmap allows the user to manually specify the arrangement of the data dimensions by defining the function arguments “x_plot”, “y_plot”, “x_facet” and “y_facet”. If not set manually, the function determines the optimal layout by analyzing the lengths of the data dimensions.

2.5 Accessibility

One of the primary objectives of the piamValidation package is to enable broad application of the tool and support IAM capacity building for developers and users of scenarios. To achieve this, the tool is published under an open-source license on GitHub and designed with user-friendliness as a central design criterion. For basic usage, a validation process can be executed with a single command, ensuring accessibility and ease of use for a broad range of users.

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-l01

Listing 1Install instructions for piamValidation.

Download

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-l02

Listing 2Command to render a validation report.

Download

For the first application, users can follow four straightforward steps: First, ensure a working R environment. If it is not already installed, one possibility could be to install R and the integrated development environment RStudio (download the freeware here: R https://www.r-project.org/, last access: 5 December 2025, and RStudio https://posit.co/products/open-source/rstudio/, last access: 5 December 2025).

Second, install the piamValidation package by running the following R command (see Listing 1).

The third step consists of the data preparation. Verify that the IAM data that needs to be validated, as well as the reference data, follow the IAMC standard format. In addition, prepare the configuration file by defining the validation checks to be performed as described in Sect. 2.2.1 and 2.2.2.

In the final step, the following R command must be executed to generate the HTML validation report, including interactive heat maps (see Listing 2). This command consolidates the file paths for the IAM data to be validated, the reference data, and the appropriate configuration file into a single operation.

For advanced users, the post-processing and visualization can be modified according to individual validation requirements by using the R data.frame containing the validation results directly. In addition, we encourage users to participate in ongoing discussions and package developments on GitHub: https://github.com/pik-piam/piamValidation/issues (last access: 5 December 2025).

3 Application case

We examine two application examples. The first, based on NGFS scenarios, emphasizing how piamValidation can be applied to an existing set of scenarios from different models, illustrating the range of potential analyses. The second demonstrates how piamValidation can be applied to assess and enhance the near-term realism of technology trends in the REMIND model. To support these applications, piamValidation offers two visualization options. We initially employ heat maps to obtain a preliminary indication of deviations. Complementarily, line plots offer a more granular perspective by depicting the detailed behavior of individual variables and their temporal dynamics.

3.1 NGFS scenarios

To demonstrate the methodological versatility of the piamValidation tool, we perform a series of validation exercises using the NGFS scenarios (data available here: https://doi.org/10.5281/zenodo.13989530, Richters et al.2024). This choice is motivated by two central considerations. First, these scenarios are not only widely recognized within the integrated assessment modeling and academic communities but are also extensively used by policymakers, financial supervisors, and other stakeholders to assess climate-related risks and transition pathways. Second, the scenarios are developed with multiple IAMs (MESSAGE3, GCAM4 and REMIND5), making them particularly suitable for illustrating the ability of piamValidation to address heterogeneous modeling frameworks.

This validation exercise is a qualitative one, focusing on the type of checks performed rather than the exact selection of threshold values. This implies that threshold exceedance does not indicate limitations in the scenario data but rather illustrate how the tool can be used to identify specific patterns. Furthermore, the plots use additional colors to indicate whether upper or lower thresholds are crossed via the function argument extraColors = TRUE when calling validateScenarios. The corresponding validation configuration file for these application cases and the markdown file to create the plots below are available via GitHub (https://github.com/pik-piam/piamValidation/blob/main/inst/config/validationConfig_publication_NGFS.csv, last access: 5 December 2025).

3.1.1 Validation overview for multiple models

As a primary application type, the piamValidation tool can be used to validate multiple variables of different models against historical reference data and near-term estimates.

The heat map overview (Fig. 4a) allows a quick identification of areas of concern. In this case, one conclusion that can be drawn from this visualization is that the near-term dynamics of wind capacities are consistently flagged across all models: while GCAM tends to overestimate the 2025 data point across all scenarios, MESSAGE and REMIND seem to rather underestimate the expansion of wind until 2030, especially in scenarios with weaker climate ambition. Using the piamValidation plotting function for time series lineplotThreshold, Fig. 4b provides a more detailed assessment of the variable.

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-f04

Figure 4Example of a validation exercise that starts by looking at (a) data of multiple variables, models, scenarios, and periods via validation heat map while reducing the region dimension to one element (“World”). After identifying a variable of interest, (b) takes a closer look at wind capacities by performing a line plot of scenario data alongside historical reference data and the validation thresholds as colored funnels.

Download

3.1.2 Model intercomparison

The piamValidation package also allows for model intercomparison exercises by selecting one model as the reference (here: MESSAGE). In this example, we examine global CO2 emissions and identify occurrences of REMIND or GCAM deviating more than 20 % (weak threshold) or 40 % (strong threshold) from MESSAGE within each scenario. The heat map in Fig. 5a reveals that the strongest deviations appear after 2050, with REMIND and GCAM showing lower emissions than MESSAGE. However, as emissions drop closer to zero, the relative differences being used as thresholds make up smaller absolute values. This becomes clearer when looking at a line plot of a specific scenario (here: “Delayed transition”) and seeing a “closing” funnel (see Fig. 5b).

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-f05

Figure 5Model intercomparison example showing (a) a heat map of the NGFS data set looking at relative deviations of CO2 emissions from REMIND and GCAM compared to MESSAGE next to (b) the line plot for the Delayed Transition scenario with a funnel of ± 20 % (green area) and 40 % (yellow/cyan area) around the MESSAGE data.

Download

Users who want to avoid this effect can choose the “difference” metric instead of the “relative” one to define constant thresholds around the reference model. Applying a buffer of ± 5/10 GtCO2 yr−1 results in a validation outcome as shown in Fig. 6.

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-f06

Figure 6Model intercomparison example now showing (a) a heat map of the NGFS data set looking at constant deviations to MESSAGE CO2 emissions and (b) the line plot for the Delayed Transition scenario with a funnel of fixed width of ±5 GtCO2 yr−1 (green area) and ± 10 GtCO2 yr−1 (yellow/cyan area) around the MESSAGE data.

Download

3.1.3 Scenario intercomparison

Similarly, scenarios can also be compared with each other (here: the reference is the “Below 2C” scenario). Consistent with the underlying scenario narratives, more ambitious scenarios such as “Net Zero 2050” and “Low Demand” are characterized by lower CO2 emissions, whereas less ambitious scenarios such as “Current Policies” and “Fragmented World” exhibit higher CO2 emissions. This application case as shown in Fig. 7 can serve as a straightforward means of conducting a preliminary plausibility check of scenario narratives.

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-f07

Figure 7Scenario intercomparison example with (a) a heat map showing which scenarios are higher or lower in emissions compared to the “Below 2C” scenario for each model and (b) a line plot for REMIND only, with constant width funnels of ±5/10 GtCO2 yr−1 around the reference scenario.

Download

3.1.4 Period intercomparison

Finally, periods can also be evaluated in relation to one another. This example checks how emissions of the periods 2025, 2030, and 2035 compare to the 2020 period. This approach can be used to validate that emissions in ambitious climate policy scenarios should decline within a plausible reduction rate. In Fig. 8 upper thresholds are 0 % and +10 % for all validated periods, while lower thresholds go from −15 % to −30 % (weaker thresholds) and −30 % to −60 % (stronger thresholds) from 2025 to 2035. As all scenarios start at the same 2020 emissions value, their funnels are also identical, which allows them to be plotted together. Note that this case also demonstrates the option of choosing asymmetrical thresholds. The results indicate that in the net zero and the low demand scenario, emission reductions are considerably more pronounced in the REMIND model compared to GCAM or MESSAGE.

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-f08

Figure 8Period intercomparison example demonstrating a feasibility study on how fast global emissions can be reduced compared to 2020, using (a) a heat map for all models and (b) a line plot for REMIND.

Download

3.2 REMIND Validate short-term technology trends

This section presents an application of the piamValidation tool to the Integrated Assessment Model REMIND (see Baumstark et al.2021 for a detailed description). To illustrate the validation process, we validate model scenarios against historical data and observational or third-party technology trend data until 2030, focusing on three key technology trends: capacity of CO2 transport and storage (CTS), sales and stock of battery electric vehicle (BEV), and capacity of offshore wind power. These technology trajectories represent some of the key energy system transformations that Luderer et al. (2022) describe: the potential for carbon management, the electrification of end-use sectors, and the composition of the electricity mix.

In scenarios of the REMIND 3.2.0 release (Luderer et al.2023), all three technologies exhibit significant deviations compared to historical and reference data. Upon identifying significant deviations, the focus shifts to improving the model and refining its underlying assumptions. The key strategies for achieving these improvements can be summarized as follows:

  1. Updating the model input data;

  2. Validating socio-economic assumptions to ensure alignment with observed trends;

  3. Revising and verifying associated cost estimates;

  4. Revising the representation of the inertia in up-scaling new technologies.

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-f09

Figure 9piamValidation heat maps of selected technology trends. Each heat map presents the performance of REMIND 3.2.0 (top) compared to the updated version, REMIND 3.5.0 (bottom) for the two climate policy scenarios: National Policies implemented (NPi, left) and “well below” 2 °C (2C, right). For BEV we validate historical modeling values against IEA Global Electric Vehicle Outlook (GEVO) historical data. For the outlook, the modeling values are validated against the IEA GEVO Announced Pledges Scenario (APS) and Stated Policies Scenario (STEPS). Similarly, wind offshore capacity is historically compared to International Renewable Energy Agency (IRENA) data, and the outlook validated Global Wind Energy Council (GWEC) data. Carbon management capacity is validated against the IEA CCUS Projects Database. An overview of all thresholds is provided in the validation configuration file (https://github.com/pik-piam/piamValidation, last access: 5 December 2025). In addition, a detailed description and community discussion on the thresholds can be accessed here: https://github.com/pik-piam/mrremind/discussions/544 (last access: 5 December 2025). In general, regional bounds are set more generously than global bounds, reflecting higher uncertainties and potential variations in accounting methods

Download

All adjustments have been incorporated into the REMIND 3.5.0 release (Luderer et al.2025). We select two climate policy scenarios spanning a wide range of policy ambition: National Policies implemented (NPi) and “well below” 2 °C (66 % likelihood) of temperature rise by the end of the century (2C)6. Both scenarios follow the Shared Socioeconomic Pathways 2 (SSP2) called Middle of the Road. The 2 °C stylized climate policy scenario assumes a peak budget of 1150 and 1000 GtCO2, respectively on total CO2 emissions from 2020 to 2100. Figure 9 illustrates how the model enhancements improve the near-term realism measured as performance against the reference data within the piamValidation tool.

More dynamic insights on model to trend deviation are presented in Fig. 10. The key developments in the updated model are described below.

https://gmd.copernicus.org/articles/18/9897/2025/gmd-18-9897-2025-f10

Figure 10Historical and near-term realism. The REMIND scenarios and thresholds presented are identical to those in Fig. 9. Scenarios updated refer to the improvements in REMIND 3.5.0

Download

3.2.1 CO2 transport and storage

CO2 transport and storage comprise the capture, transport, and storage or use of CO2 from different CO2 sources. All CO2 that is captured, transported, and stored is accounted for in the variable “CO2|Transport and Storage”. The carbon transport and storage infrastructure is represented as a single technology in REMIND. The construction of geologic carbon storage infrastructure is associated with long lead times, as the development of geologic storage typically takes 7 to 10 years from exploration to industrial-scale injection (Bui et al.2018). Upscaling may be further limited by the availability of geological engineers and drilling rigs (Budinis et al.2018). In addition, a wide range of cost estimates have been documented. While studies on technical cost range between 3 and 20 USD (t CO2)−1 (Budinis et al.2018; IPCC2014), costs announced for projects reach up to 80 USD (t CO2)−1 (Bellona2022; Jakobsen et al.2017; IOGP2023).

There are thus two trends relevant for model improvement. First, storage potentials until 2030 are limited by projects announced today due to the long lead times. We thus refer to the database on Carbon Capture, Utilization, and Storage (CCUS) projects, published by the International Energy Agency (IEA) in 2023 that is also used for the validation reference values. To improve the near-term realism of geologic storage in the model, we introduce regional lower and upper bounds. The lower bounds in 2025 and 2030 are based on the announced capacities of storage projects that are either operational or under construction (with announced starts of operation before 2025 and 2030, respectively). The upper bounds for 2025 and 2030 additionally include 40 % of the capacity of all announced projects that report a start of operation by 2025 or 2030, respectively. Note that for these, the final investment decision is often pending, thus a certain share of unrealized project capacities is assumed based on historic delay and failure rates. Furthermore, as the EU27 pursues a transnational, Europe-wide CCS infrastructure, the capacity of a given CCS project is not fully attributed to the respective member state within the model. Instead, the EU27 CCS capacities are pooled (including 50 % of Norway's and 50 % of UKs near-term potential) and redistributed across member states based on GDP. The second improvement concerns the costs of carbon transport and storage. The technical cost estimate in REMIND 3.2.0 of 7.5 USD (t CO2)−1 appears too low, despite REMIND modeling the cost of upscaling through adjustment cost. We thus increase our capital cost estimate to reach 12 USD (t CO2)−1 informed by the average in Budinis et al. (2018).

3.2.2 Electric vehicle stock and sales

REMIND’s transportation sector is modeled by the simulation model “Energy Demand Generator-Transport” (EDGE-T) (Rottoli et al.2021) that is integrated into the broader sociotechnical framework of REMIND. EDGE-T simulates technology and transport mode decisions relying on a multinomial logit approach. The decision tree, the logit, comprises a nested structure of subsequent choices among comparable alternatives for transport modes and technologies. This established approach is suited for the modeling of end-consumer decisions under probabilistic conditions and allows for including non-economic factors, such as inconvenience costs, in the decision simulation.

In the transport sector, a prominent example is the increase in electric vehicles (EVs), which is not mainly driven by CO2 prices and, consequently, fuel costs, but rather by technical innovations leading to up-front investment cost reductions, policy incentives, and shifts in how consumers assess this technology alternative. The piamValidation tool revealed deviations from near-term estimates for stocks and sales of EVs. Model parameters were thus updated to reflect technological advancement and better understanding of the uptake dynamics of electric vehicles. Building on the Global EV Outlook (IEA2024) and its information about the shares of electric vehicles in car sales and car fleets in different world regions, the model representation of consumer attitudes (so-called inconvenience costs) towards electromobility was updated. Furthermore, input data processing was critically assessed, e.g., by screening for implausible outliers in input data, and correction routines were implemented. As high-quality transport data, such as energy service demand, vehicle fleet characteristics, and vehicle costs, are scarce, verifying data plausibility across countries, technologies, and vehicle sizes is crucial for improving model and scenario quality.

3.2.3 Offshore wind capacity

The competition between onshore and offshore wind power is largely determined by factors like grid integration costs or the matching between generation and demand. REMIND 3.2.0 – like other long-term global integrated assessment models – represents these drivers in a very aggregated and parameterized way, since it does not explicitly resolve the relevant spatial and temporal scales. Consequently, the model limits the share of offshore wind generation in total wind energy to a certain range. This share is constrained in the near-term to reflect the historical lag between onshore and offshore deployment.

In REMIND 3.2.0, historical capacities are available until 2020, at which point offshore wind is still a nascent technology with around 35 GW of installed capacity worldwide. To prevent the model from deploying offshore earlier than observed in the real-world, only a portion of the total offshore wind potential is available: before 2010, none of the potential is available, then the available portion increases gradually until 2050. Conversely, specific equations ensure that the development of offshore capacity is not too slow compared to onshore after 2025: in a given region, if the available offshore potential is four times as high as onshore, the equations constrain the model to deploy each year at least twice as much offshore as onshore new capacity.

The piamValidation tool has revealed that these constraints were overly pessimistic: wind offshore deployment in REMIND 3.2.0 is slower than observed in real data, particularly in Europe. To address this discrepancy, we updated the model to match the real data more closely. REMIND 3.5.0 uses newly released data by IRENA (2024b) to compute consolidated historical capacities for 2020 and a lower bound for 2025 based on 2023 capacity. Moreover, the model allows for an accelerated phase-in of offshore wind by making the full potential available by 2030 instead of 2050. The equations linking offshore to onshore are maintained after 2030 in order to follow the rising trend of offshore wind despite its higher investment costs.

4 Conclusions

Integrated Assessment Models are central to the climate negotiations and play a key role in the IPCC assessments and thus at the interface between climate science and policy. However, they face challenges and criticism regarding their transparency (Robertson2021), their ability to capture technology trends (Keppo et al.2021; Anderson2019) and their overall reliability (Wilson et al.2021).

We introduce the open-source R package piamValidation, a versatile tool designed to address multiple use cases in the validation of IAM scenario data. This package facilitates validation across IAM scenarios, models, and temporal periods. Furthermore, it supports validation against historical, observational, and benchmark datasets, using time series, thresholds, or growth rates as a reference. To enhance reliability, piamValidation can be seamlessly integrated into the regular development workflow of IAMs, enabling comprehensive sanity checks and performance evaluations.

We outline the structure of the piamValidation package, emphasizing its user-friendly design to facilitate widespread adoption within the IAM community and further stakeholders. The aim is to enhance the overall validity and reliability of IAMs through the systematic application of the tool. For this purpose, we provide a structured guideline together with standard use cases, demonstrating the application of the tool through single-command execution.

We demonstrate the versatility of the piamValidation tool in two complementary application cases. They highlight how the tool can be used broadly across contexts as well as to refine the near-term dynamics of technology pathways in REMIND. The REMIND case study demonstrates that the usefulness of the tool for historical and benchmark validation critically depends on the chosen validation settings and the availability of reliable reference data. Consequently, the utility of the piamValidation tool would be enhanced by broader adoption and the systematic documentation of configuration files within the IAM community.

The primary benefit of the piamValidation tool lies in its ability to efficiently provide an overview of variable deviations within large-scale datasets. It also facilitates systematic validation across IAM scenarios and supports sanity checks throughout model development processes. Furthermore, the tool serves as a potential platform for the IAM community to collaboratively build a repository of knowledge on reasonable validation thresholds.

Despite its utility, the piamValidation tool has some important limitations that fall into three broad areas. First, the quality of any validation exercise cannot be guaranteed by the tool itself but instead depends on user and community choices. Most importantly, the identification of meaningful validation cases, the selection of appropriate reference data, and the definition of reasonable thresholds all substantially influence the outcome of the validation exercise. Although these challenges are not technical limitations of the piamValidation tool, they substantially influence the quality, consistency, and acceptance of the validation results. Second, certain limitations arise from the design of the tool. In particular, caution is required when thresholds are defined in terms of relative deviations and the validation values approach zero. Under such conditions, even minimal absolute differences can manifest as disproportionately large relative deviations, complicating the interpretation of results. In these cases, the use of absolute deviations is preferable. Finally, the tool is constrained by challenges in data management. Harmonizing scenario data with reference sources often requires substantial effort to ensure consistency in units, definitions, sectoral coverage, and technological detail. This integration process can be time-consuming.

Ongoing improvements are being made to the piamValidation tool, with continuous development efforts focused on enhancing transparency of error handling, unit testing of various use cases and additional capabilities. These include three central development areas: First, we combine relative and absolute metrics to address scale dependency in relative deviations, particularly near zero. To reduce this scale dependency and increase the user-friendliness of the piamValidation tool, we are working on a new feature that includes a general tolerance. We plan to implement an optional lump-sum buffer, defined as a fraction of the largest reference value, within the processing workflow. Second, we plan to enable differentiation across validation criteria, such as feasibility, sustainability, or equity. A third key development area centers around the ability to handle, compare and vet scenarios from an extensive scenario database. These improvements are designed to make the tool more user-friendly and versatile for the IAM community.

In addition to these development areas, we anticipate a process of continuous improvement for the validation of realism in technology trends. Central to this is the ongoing expansion of reference variables and the refinement of existing thresholds, which requires regularly updating current data sources while monitoring and integrating new ones. Further developments may include extending validation metrics to assess mid-term feasibility and policy robustness. The extent and pace of these improvements will be accelerated by the tool's widespread application, particularly through the sharing of configuration files and active participation in GitHub discussions (https://github.com/pik-piam/mrremind/discussions, last access: 5 December 2025), which facilitate feedback, highlight priorities, and guide future enhancements.

Code and data availability

The REMIND-MAgPIE code is implemented in GAMS, whereas code and data management is done using R. The REMIND 3.2.0 code is archived at Zenodo (https://doi.org/10.5281/zenodo.7852740, Luderer et al.2023) and REMIND 3.5.0 at (https://doi.org/10.5281/zenodo.15147820, Luderer et al.2025). The technical model documentation is available on the common integrated assessment model documentation website (https://www.iamcdocumentation.eu/Model_Documentation_-_REMIND-MAgPIE, last access: 5 December 2025), and is published as open-source code (https://github.com/remindmodel/remind (last access: 5 December 2025) under the GNU Affero General Public License v3.0).

A repository for the source code of the piamValidation package is available via GitHub at https://github.com/pik-piam/piamValidation (last access: 5 December 2025) under the GNU Lesser General Public License v3.0. The piamValidation code is archived at Zenodo under (https://doi.org/10.5281/zenodo.17661999, Weigmann et al.2025).

The reproduction of the plots in this publication is documented at https://pik-piam.github.io/piamValidation/articles/publication.html (last access: 5 December 2025) and can be achieved with the RMarkdown files at https://github.com/pik-piam/piamValidation/blob/main/inst/markdown/validationReport_publication.Rmd (last access: 5 December 2025) and https://github.com/pik-piam/piamValidation/blob/main/inst/markdown/validationReport_publication_NGFS.Rmd (last access: 5 December 2025). The validation in this study is performed with the configuration files at https://github.com/pik-piam/piamValidation/blob/main/inst/config/validationConfig_publication.csv (last access: 5 December 2025) and https://github.com/pik-piam/piamValidation/blob/main/inst/config/validationConfig_publication_NGFS.csv (last access: 5 December 2025).

The NGFS v5.0 data set has been published via Zenodo under a Public License and is available via the DOI https://doi.org/10.5281/zenodo.13989530, Richters et al. (2024).

Author contributions

GL, LB and EK supervised the development of the piamValidation package. The piamValidation package is initiated and led by PW, where FL and OR contributed to its development. FL, AM, TD, JH, JM, RP, OR, LV, NB, FB, CC, RM and GL contributed to the REMIND model development. All authors commented, reviewed and contributed to the final manuscript.

Competing interests

At least one of the (co-)authors is a member of the editorial board of Geoscientific Model Development. The peer-review process was guided by an independent editor, and the authors also have no other competing interests to declare.

Disclaimer

Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors. Views expressed in the text are those of the authors and do not necessarily reflect the views of the publisher.

Acknowledgements

The authors are grateful to the attendees of the Seventeenth Assessment Modeling Consortium (IAMC) Annual Meeting 2024 for their valuable feedback and reflections. This work received support from the European Commission Directorate-General of Climate Action (DG CLIMA) under Service Contract No. 14020241/2022/884157/SER/CLIMA.A.2 CLIMA/2022/EA-RP/0007 and the Horizon Europe Project PRISMA, under the grant agreement number 1010816.

Financial support

This research has been supported by the Directorate-General for Climate Action (grant no. 14020241/2022/884157/SER/CLIMA.A.2 CLIMA/2022/EA-RP/0007) and the European Climate, Infrastructure and Environment Executive Agency (grant no. 1010816).

The article processing charges for this open-access publication were covered by the Potsdam Institute for Climate Impact Research (PIK).

Review statement

This paper was edited by Yongze Song and reviewed by Gianni Bellocchi, Béatrice Cointe, and two anonymous referees.

References

Anderson, K.: Wrong tool for the job: Debating the bedrock of climate-change mitigation scenarios, Nature, 573, 348–348, 2019. a

Baumstark, L., Bauer, N., Benke, F., Bertram, C., Bi, S., Gong, C. C., Dietrich, J. P., Dirnaichner, A., Giannousakis, A., Hilaire, J., Klein, D., Koch, J., Leimbach, M., Levesque, A., Madeddu, S., Malik, A., Merfort, A., Merfort, L., Odenweller, A., Pehl, M., Pietzcker, R. C., Piontek, F., Rauner, S., Rodrigues, R., Rottoli, M., Schreyer, F., Schultes, A., Soergel, B., Soergel, D., Strefler, J., Ueckerdt, F., Kriegler, E., and Luderer, G.: REMIND2.1: transformation and innovation dynamics of the energy-economic system within climate and sustainability limits, Geosci. Model Dev., 14, 6571–6603, https://doi.org/10.5194/gmd-14-6571-2021, 2021. a, b

Beck, S. and Oomen, J.: Imagining the corridor of climate mitigation – What is at stake in IPCC's politics of anticipation?, Environmental Science & Policy, 123, 169–178, 2021. a

Bellona: Norway's Longship CCS project, https://network.bellona.org/content/uploads/sites/3/2020/10/Longship-Briefing_Bellona-1.pdf (last access: 5 December 2025), 2022. a

Bertram, C., Brutschin, E., Drouet, L., Luderer, G., van Ruijven, B., Aleluia Reis, L., Baptista, L. B., de Boer, H.-S., Cui, R., Daioglou, V., Fosse, F., Fragkiadakis, D., Fricko, O., Fujimori, S., Hultman, N., Iyer, G., Keramidas, K., Krey, V., Kriegler, E., Lamboll, R. D., Mandaroux, R., Rochedo, P., Rogelj, J., Schaeffer, R., Silva, D., Tagomori, I., van Vuuren, D., Vrontisi, Z., and Riahi, K.: Feasibility of peak temperature targets in light of institutional constraints, Nature Climate Change, 14, 954–960, 2024. a

Brutschin, E., Pianta, S., Tavoni, M., Riahi, K., Bosetti, V., Marangoni, G., and Van Ruijven, B. J.: A multidimensional feasibility evaluation of low-carbon scenarios, Environmental Research Letters, 16, 064069, https://doi.org/10.1088/1748-9326/abf0ce, 2021. a

Budinis, S., Krevor, S., Mac Dowell, N., Brandon, N., and Hawkes, A.: An assessment of CCS costs, barriers and potential, Energy strategy reviews, 22, 61–81, 2018. a, b, c

Bui, M., Adjiman, C. S., Bardow, A., Anthony, E. J., Boston, A., Brown, S., Fennell, P. S., Fuss, S., Galindo, A., Hackett, L. A., Hallett, J. P., Herzog, H. J., Jackson, G., Kemper, J., Krevor, S., Maitland, G. C., Matuszewski, M., Metcalfe, I. S., Petit, C., Puxty, G., Reimer, J., Reiner, D. M., Rubin, E. S., Scott, S. A., Shah, N., Smit, B., Trusler, J. P. M., Webley, P., Wilcox, J., and Mac Dowell, N.: Carbon capture and storage (CCS): the way forward, Energy & Environmental Science, 11, 1062–1176, 2018. a

Carbon Brief: Just 15 countries account for 98 % of new coal-power development, https://www.carbonbrief.org/guest-post-just-15-countries-account-for-98-of-new-coal-power-development/ (last access: 10 October 2024), 2024. a

Cointe, B., Cassen, C., and Nadaï, A.: Organising policy-relevant knowledge for climate action: integrated assessment modelling, the IPCC, and the emergence of a collective expertise on socioeconomic emission scenarios, Science & Technology Studies, https://doi.org/10.23987/sts.65031, 2019. a

Creutzig, F., Hilaire, J., Nemet, G., Müller-Hansen, F., and Minx, J. C.: Technological innovation enables low cost climate change mitigation, Energy Research & Social Science, 105, 103276, https://doi.org/10.1016/j.erss.2023.103276, 2023. a

Edenhofer, O. and Minx, J.: Mapmakers and navigators, facts and values, Science, 345, 37–38, 2014. a

Giannousakis, A., Hilaire, J., Nemet, G. F., Luderer, G., Pietzcker, R. C., Rodrigues, R., Baumstark, L., and Kriegler, E.: How uncertainty in technology costs and carbon dioxide removal availability affect climate mitigation pathways, Energy, 216, 119253, https://doi.org/10.1016/j.energy.2020.119253, 2021. a

IEA: Global EV Outlook 2024, licence: CC BY 4.0, https://www.iea.org/reports/global-ev-outlook-2024 (last access: 5 December 2025), 2024. a

IOGP: Creating a sustainable business case for CCS value chains – the needed funding and de-risking mechanisms, https://iogpeurope.org/wp-content/uploads/2023/11/Creating-a-Business-Case-for-CCS-Value-Chains-IOGP-Europe.pdf (last access: 5 December 2025), 2023. a

IPCC: IPCC fifth assessment report – synthesis report, IPPC Rome, Italy, https://www.ipcc.ch/site/assets/uploads/2018/02/SYR_AR5_FINAL_full.pdf last access: 5 December 2025, 2014. a

IRENA: Renewable energy highlights, https://www.irena.org/-/media/Files/IRENA/Agency/Publication/2023/Jul/Renewable_energy_highlights_July_2023.pdf (last access: 10 October 2024), 2024a. a

IRENA: Renewable capacity statistics, Tech. rep., International Renewable Energy Agency, https://www.irena.org/-/media/Files/IRENA/Agency/Publication/2024/Mar/IRENA_RE_Capacity_Statistics_2024.pdf (last access: 5 December 2025), 2024b. a

Jakobsen, J., Roussanaly, S., and Anantharaman, R.: A techno-economic case study of CO2 capture, transport and storage chain from a cement plant in Norway, Journal of cleaner production, 144, 523–539, 2017. a

Keppo, I., Butnar, I., Bauer, N., Caspani, M., Edelenbosch, O., Emmerling, J., Fragkos, P., Guivarch, C., Harmsen, M., Lefèvre, J., Le Gallic, T., Leimbach, M., McDowall, W., Mercure, J.-F., Schaeffer, R., Trutnevyte, E., and Wagner, F.: Exploring the possibility space: Taking stock of the diverse capabilities and gaps in integrated assessment models, Environmental Research Letters, 16, 053006, https://doi.org/10.1088/1748-9326/abe5d8, 2021. a, b

Krey, V., Guo, F., Kolp, P., Zhou, W., Schaeffer, R., Awasthy, A., Bertram, C., de Boer, H.-S., Fragkos, P., Fujimori, S., He, C., Iyer, G., Keramidas, K., Köberle, A. C., Oshiro, K., Reis, L. A., Shoai-Tehrani, B., Vishwanathan, S., Capros, P., Drouet, L., Edmonds, J. E., Garg, A., Gernaat, D. E., Jiang, K., Kannavou, M., Kitous, A., Kriegler, E., Luderer, G., Mathur, R., Muratori, M., Sano, F., and van Vuuren, D. P.: Looking under the hood: a comparison of techno-economic assumptions across national and global integrated assessment models, Energy, 172, 1254–1267, 2019. a, b

Luderer, G., Madeddu, S., Merfort, L., Ueckerdt, F., Pehl, M., Pietzcker, R., Rottoli, M., Schreyer, F., Bauer, N., Baumstark, L., Bertram, C., Dirnaichner, A., Humpenöder, F., Levesque, A., Popp, A., Rodrigues, R., Strefler, J., and Kriegler, E.: Impact of declining renewable energy costs on electrification in low-emission scenarios, Nature Energy, 7, 32–42, 2022. a

Luderer, G., Bauer, N., Baumstark, L., Bertram, C., Leimbach, M., Pietzcker, R., Strefler, J., Aboumahboub, T., Abrahão, G., Auer, C., Benke, F., Bi, S., Dietrich, J., Dirnaichner, A., Fuchs, S., Führlich, P., Giannousakis, A., Gong, C. C., Haller, M., Hasse, R., Hilaire, J., Hoppe, J., Klein, D., Koch, J., Körner, A., Kowalczyk, K., Kriegler, E., Levesque, A., Lorenz, A., Ludig, S., Lüken, M., Malik, A., Mandaroux, R., Manger, S., Merfort, A., Merfort, L., Moreno-Leiva, S., Mouratiadou, I., Odenweller, A., Pehl, M., Pflüger, M., Piontek, F., Popin, L., Rauner, S., Richters, O., Rodrigues, R., Roming, N., Rottoli, M., Schmidt, E., Schötz, C., Schreyer, F., Schultes, A., Sörgel, B., Ueckerdt, F., Verpoort, P., and Weigmann, P.: REMIND – REgional Model of INvestments and Development 3.2.0, Zenodo [code], https://doi.org/10.5281/zenodo.7852740, 2023. a, b

Luderer, G., Bauer, N., Baumstark, L., Bertram, C., Leimbach, M., Pietzcker, R., Strefler, J., Aboumahboub, T., Abrahão, G., Auer, C., Bantje, D., Beier, F., Benke, F., Bi, S., Dietrich, J. P., Dirnaichner, A., Dorndorf, T., Duerrwaechter, J., Fuchs, S., Führlich, P., Giannousakis, A., Gong, C. C., Hagen, A., Haller, M., Hasse, R., Hayez, L., Hilaire, J., Hofbauer, V., Hoppe, J., Klein, D., Koch, J., Köhler-Schindler, L., Körner, A., Kowalczyk, K., Kriegler, E., Lécuyer, F., Levesque, A., Lorenz, A., Ludig, S., Lüken, M., Malik, A., Mandaroux, R., Manger, S., Merfort, A., Merfort, L., Moreno-Leiva, S., Mouratiadou, I., Müßel, J., Odenweller, A., Pehl, M., Pflüger, M., Piontek, F., Popin, L., Rauner, S., Richters, O., Rodrigues, R., Roming, N., Rosemann, R., Rottoli, M., Rüter, T., Salzwedel, R., Sauer, P., Schmidt, E., Schötz, C., Schreyer, F., Schultes, A., Sitarz, J., Sörgel, B., Ueckerdt, F., Verpoort, P., Weigmann, P., and Weiss, B.: REMIND – REgional Model of INvestments and Development 3.5.0, Zenodo [code], https://doi.org/10.5281/zenodo.15147820, 2025.  a, b

Richters, O., Kriegler, E., Bertram, C., Cui, R., Edmonds, J., Fawcett, A., Fuhrman, J., George, M., Hackstock, P., Hurst, I., Ju, Y., Kotz, M., Liadze, I., Min, J., Piontek, F., Sanchez Juanino, P., Sferra, F., Stevanovic, M., van Ruijven, B., Weigmann, P., Wenz, L., Westphal, M. I., Zwerling, M., Abrahão, G., Baumstark, L., Bresch, D. N., Chen, D. M.-C., Dietrich, J. P., Durga, S., Fricko, O., Hasse, R., Hoppe, J., Humpenöder, F., Iyer, G., Javaid, A., Joshi, S., Kikstra, J., Kishimoto, P., Klein, D., Koch, J., Krey, V., Kropf, C. M., Lewis, J., Lochner, E., Luderer, G., Maczek, F., Mandaroux, R., Mastrucci, A., Meinshausen, M., Meng, M., Merfort, A., Nicholls, Z., Patel, P., Pehl, M., Pelz, S., Popp, A., Rüter, T., Sauer, I., Schreyer, F., Ünlü, G., von Jeetze, P., Zhao, A., and Zhao, X.: NGFS climate scenarios data set, Zenodo [data set], https://doi.org/10.5281/zenodo.13989530, 2024. a

Robertson, S.: Transparency, trust, and integrated assessment models: An ethical consideration for the Intergovernmental Panel on Climate Change, Wiley Interdisciplinary Reviews: Climate Change, 12, e679, https://doi.org/10.1002/wcc.679, 2021. a

Rottoli, M., Dirnaichner, A., Kyle, P., Baumstark, L., Pietzcker, R., and Luderer, G.: Coupling a detailed transport model to the integrated assessment model REMIND, Environmental Modeling & Assessment, 26, 891–909, 2021. a

Stanton, E. A., Ackerman, F., and Kartha, S.: Inside the integrated assessment models: Four issues in climate economics, Climate and Development, 1, 166–184, 2009. a

Van Beek, L., Hajer, M., Pelzer, P., van Vuuren, D., and Cassen, C.: Anticipating futures through models: the rise of Integrated Assessment Modelling in the climate science-policy interface since 1970, Global Environmental Change, 65, 102191, https://doi.org/10.1016/j.gloenvcha.2020.102191, 2020. a

Weigmann, P., Richters, O., Lécuyer, F., and Koch, J.: piamValidation: Validation Tools for PIK-PIAM, Zenodo [code], https://doi.org/10.5281/zenodo.17661999, 2025. a

Wilson, C., Guivarch, C., Kriegler, E., Van Ruijven, B., Van Vuuren, D. P., Krey, V., Schwanitz, V. J., and Thompson, E. L.: Evaluating process-based integrated assessment models of climate change mitigation, Climatic Change, 166, 1–22, 2021. a, b

1

The practice of benchmarking models, periods, or scenarios against one specific element of the respective dimension is referred to as intercomparison throughout the remainder of this text.

2

Note: The same holds for ref_scenario and ref_period.

3

MESSAGEix-GLOBIOM: Model for Energy Supply Strategy Alternatives and their General Environmental Impact – GLObal BIOsphere Model.

4

GCAM: Global Change Analysis Model.

5

REMIND-MAgPIE: REgional Model of INvestments and Development – The Model of Agricultural Production and its Impact on the Environment.

6

The exact original scenario names are “SSP2EU-NPi” and “SSP2EU-PkBudg1150” in REMIND 3.2.0 and “SSP2-NPi2025” and “SSP2-PkBudg1000” in REMIND 3.5.0 for NPi and 2C respectively. They have been renamed for clarity and brevity.

Download
Short summary
We present the Potsdam Integrated Assessment Modeling validation tool, piamValidation, an open-source R package for validating climate scenarios. The tool enables comparison of model outputs with historical data, feasibility constraints and alternative scenarios. Designed as a community resource, validation configuration files can serve as a knowledge-sharing platform. The main objective is to improve the credibility of Integrated Assessment Models by promoting standardized validation practices.
Share