Articles | Volume 13, issue 8
https://doi.org/10.5194/gmd-13-3627-2020
https://doi.org/10.5194/gmd-13-3627-2020
Methods for assessment of models
 | 
21 Aug 2020
Methods for assessment of models |  | 21 Aug 2020

Evaluating simulated climate patterns from the CMIP archives using satellite and reanalysis datasets using the Climate Model Assessment Tool (CMATv1)

John T. Fasullo
Abstract

An objective approach is presented for scoring coupled climate simulations through an evaluation against satellite and reanalysis datasets during the satellite era (i.e., since 1979). The approach is motivated, described, and applied to available Coupled Model Intercomparison Project (CMIP) archives and the Community Earth System Model (CESM) Version 1 Large Ensemble archives with the goal of robustly benchmarking model performance and its evolution across CMIP generations. A scoring system is employed that minimizes sensitivity to internal variability, external forcings, and model tuning. Scores are based on pattern correlations of the simulated mean state, seasonal contrasts, and ENSO teleconnections. A broad range of feedback-relevant fields is considered and summarized on discrete timescales (climatology, seasonal, interannual) and physical realms (energy budget, water cycle, dynamics). Fields are also generally chosen for which observational uncertainty is small compared to model structural differences.

Highest mean variable scores across models are reported for well-observed fields such as sea level pressure, precipitable water, and outgoing longwave radiation, while the lowest scores are reported for 500 hPa vertical velocity, net surface energy flux, and precipitation minus evaporation. The fidelity of models is found to vary widely both within and across CMIP generations. Systematic increases in model fidelity in more recent CMIP generations are identified, with the greatest improvements occurring in dynamic and energetic fields. Such examples include shortwave cloud forcing and 500 hPa eddy geopotential height and relative humidity. Improvements in ENSO scores with time are substantially greater than for climatology or seasonal timescales.

Analysis output data generated by this approach are made freely available online from a broad range of model ensembles, including the CMIP archives and various single-model large ensembles. These multimodel archives allow for an expeditious analysis of performance across a range of simulations, while the CESM large ensemble archive allows for estimation of the influence of internal variability on computed scores. The entire output archive, updated and expanded regularly, can be accessed at http://webext.cgd.ucar.edu/Multi-Case/CMAT/index.html (last access: 18 August 2020).

1 Introduction

Global climate models were first developed over half a century ago (Hunt and Manabe, 1968; Manabe et al., 1975) and have provided insight into the climate system on a range of issues including the roles of various physical processes in the climate system and the attribution of climate events. They also are key tools for near-term initialized prediction and long-term boundary forced projections. Given their relevance for addressing issues of considerable socioeconomic importance, climate models are increasingly being looked to for guiding policy-relevant decisions on long timescales and on regional levels. Many barriers exist however, and chief amongst which are the biases in climate model representations of the physical system.

Adequate evaluation of climate models is nontrivial however. A key obstacle is that the longest observational records tend to monitor temperature and sea level pressure and are therefore not directly related to many of the fields thought to govern climate variability and change, such as for example cloud radiative forcing and rainfall (Burrows et al., 2018). Global direct observations of more physically relevant fields exist but are available exclusively from satellite and thus are limited in duration, with some of the most important data records beginning only in recent decades. Over longer timescales, uncertainties in forcing external to the climate system (e.g., anthropogenic aerosols) further complicate model evaluation. Benchmarks of model performance must therefore be designed to deal with associated uncertainties and minimize their influence.

1.1 Motivations

Climate modeling centers continually refine their codes with the goal of improving their models. The Climate Model Intercomparison Project (CMIP) is an effort to systematically coordinate and release targeted climate model experiments of high interest in the science community and has thus far provided three major releases, including CMIP3 (Meehl et al., 2007), CMIP5 (Taylor et al., 2012), and CMIP6 versions (Eyring et al., 2016). Major advances have also recently been made in key observationally based climate datasets (as discussed herein). An opportunity has therefore arisen to take stock of these simulation archives and conduct a retrospective assessment of progress that has been made and challenges that remain.

While individual models are widely scrutinized, systematic surveys of model performance are relatively rare. Evaluation of single CMIP generations have been conducted, and these have been uniquely useful for identifying canonical model biases (Gleckler et al., 2008; Pincus et al., 2008). It is the goal of this study to provide a similar benchmarking of models but one considerably expanded in scope in considering multiple CMIP versions and using newly available process-relevant observations that contextualize model–observation differences with respect to both internal variability and observational uncertainty. An additional goal is to provide related diagnostic outputs directly to the community. Both the graphical and data outputs generated may potentially be incorporated into broader community packages such as ESMValTool (Eyring et al., 2020), thus providing a unique evaluation of fully coupled physical climate states that encompasses both climatological means and temporal variations, which accounts for key uncertainties, and that benchmarks models across CMIP generations.

1.2 Challenges

A number of challenges exist for efforts aimed at comprehensively assessing climate model fidelity. Observations of many fields that are central to climate variability and change (e.g., cloud microphysics, entrainment rates, aerosol–cloud interactions; Knutti et al., 2010) are not observed on the global, multidecadal timescales required to comprehensively evaluate models. Fields for which observations do exist often entail uncertainties that are large, particularly at times when the spatial sampling of observing networks is poor (e.g., SST datasets) or for fields that contain significant uncertainty in satellite-based retrieval (e.g., surface turbulent and radiative fluxes). For instances in which extended data records are unavailable, associated sensitivity to internal variability and externally imposed forcing, which also contains major uncertainties, must be considered, and evaluation of trends are particularly susceptible. In addition, model tuning methods vary widely across centers (e.g., Hourdin et al., 2017; Schmidt et al., 2017), and in instances where climate fields are explicitly tuned, direct comparison against observations is unwarranted.

1.3 Approach

The need for objective climate model analysis was highlighted in the 2010 IPCC Expert Meeting on Assessing and Combining Multi-Model Climate Projections (Knutti et al., 2010). Its synthesis report detailed a number of summary recommendations including the consideration of feedback-relevant, process-based fields and the implementation of metrics that are both simple and statistically robust. In addition, fields were recommended for which observational uncertainty and internal variability are both quantifiable and small relative to model structural differences. The reliance on any single evaluation dataset was also deemed problematic in that doing so might be both susceptible to compensating errors and insufficient to fully characterize intermodel contrasts. The approach here is guided, in part, by these recommendations.

Various model analysis efforts have focused on surface temperature (e.g., Braverman et al., 2017; Lorenz et al., 2018). A thorough evaluation of climate model thermodynamics is provided by the TheDiaTo as described in Lembo et al. (2019). Complex measures of model performance that allow for a richer comparison of model statistics against observations have also been discussed (Gibbs and Su, 2002) and proposed (Ghil, 2015). The approach adopted here highlights instead the main components of the energy and water cycles using simple diagnostic measures. Objective assessments of CMIP3 performance based on the mean climate state using these fields were performed in Gleckler et al. (2008) and Pincus et al. (2008). The goal of this work is to complement and extend these efforts in including an analysis of both the mean state and variability across three generations of CMIP simulations while distinguishing between timescales and realms of diagnostics as well as using improved observational datasets and constraints (described below). As the skill of a given climate model is likely to depend on the relevant application (Gleckler et al., 2008; Pierce et al., 2009; Knutti et al., 2017), the scores computed herein are made widely available to the community and may help guide formation of optimal model subsets for targeted applications.

The consideration of multiple CMIP generations is motivated in part by reported shifts in model behavior, such as for example the apparent increase in climate sensitivity to carbon dioxide in some models (Gettelman et al., 2019; Golaz et al., 2019; Neubauer et al., 2019). Do such shifts accompany systematic improvements in models and if so, in what fields? It is also of a more general interest to quantify canonical biases in models, their changes in successive model generations, and persistent biases affecting the most recent generations of climate models. The specific questions addressed here therefore include the following: what improvements have occurred across model generations, and what persistent biases remain? What process-relevant, well-observed fields are models most skillful in reproducing? To what extent are apparent improvements and persisting biases robustly detectible in the presence of internal climate variability, particularly as they relate to brief satellite records?

2 Methods

The analysis approach consists of computing a range of scores based on pattern correlations encompassing three climatic timescales: the climatological annual mean (annual), seasonal mean contrasts (June–July–August and December–January–February), and ENSO teleconnection patterns –computed from the 12-month July through June mean regressions against Niño3.4 (area-averaged SSTA 5 S–5 N, 170–120 W) sea surface temperatures (SST). The choice of ENSO as a model diagnostic is motivated in part by the demands involved in its accurate simulation arising from the highly coupled nature of the mode, which includes feedbacks between clouds, diabatic heating, and winds in the atmosphere, and currents and steric structure in the ocean (e.g., Cheng et al., 2019). Variables are classified according to three variable types (or realms) corresponding to the energy budget, water cycle, and dynamics. To reduce the influence of internal variability, the time period over which these fields are considered is at least 20 years, though the availability of some datasets allows for the use of longer periods, further reducing the susceptibility of the analysis to internal variability. Contemporaneous time intervals are also chosen to provide for maximum overlap between observed and simulated fields. The variables selected for consideration are chosen based on availability and judgment of their importance in simulating climate variability and change. In part this judgment is based on a recent community solicitation (Burrows et al., 2018), and some of the fields included (e.g., TOA fluxes) are deemed by experts to be optimal metrics for model evaluation (e.g., Baker and Taylor, 2016).

2.1 Observational datasets

2.1.1 The energy budget realm

Energy budget fields considered consist broadly of TOA radiative fluxes and cloud forcing, vertically integrated atmospheric energy divergence and tendency, and surface heat fluxes. Radiative fluxes at TOA are taken from the Clouds and Earth's Radiant Energy System (CERES) Energy Balance and Filled Version 4.1 dataset (EBAFv4.1; Loeb et al., 2018). The dataset offers a number of improvements over earlier versions and datasets, with improved angular distribution models and scene identification, but is perhaps most notable for its recently updated derivation of cloud radiative forcing (CF). Historically CF has been estimated from observations by differencing cloudy and neighboring clear regions, with the effect of aliasing meteorological contrasts between the regions (whereas models merely remove clouds from their radiative transfer scheme using colocated meteorology). In the EBAFv4.1, fields from NASA's GEOS-5 reanalysis (Borovikov et al., 2019) are used to estimate fluxes and CF for colocated (rather than remote) atmospheric conditions, thus providing for a more analogous comparison to models. From CERES, the TOA net shortwave (ASR), outgoing longwave (OLR), and net (RT) radiative fluxes are used. In addition, estimates of shortwave CF (SWCF) and longwave CF (LWCF) are used.

Derived from the ERA-Interim reanalysis (Dee et al., 2011), vertical integrals of atmospheric energy are used to both assess the total energy divergence within the atmosphere (∇⋅AE) and its tendency (AE/t). This provides important insight into the regional generation of atmospheric transports and their cumulative influence on the global energy budget (e.g., Fasullo and Trenberth, 2008). They are also an energy budget component necessary for computing the net surface energy fluxes from the residual of RT, ∇⋅AE , and AE/t. Given the challenges of directly observing the net surface flux, a residual method is likely the best available method for estimating the large-scale evaluation of the surface heat budget. The method has been demonstrated to achieve an accuracy on par with direct observations on regional scales and has proven superior on large scales, where the atmospheric divergences on which they rely become small, converging to zero by definition in the global mean (Trenberth and Fasullo, 2017). Uncertainty estimation of CERES fluxes is also well documented (Loeb et al., 2018).

2.1.2 The water cycle realm

Water cycle fields considered include precipitation (P), evaporation minus precipitation (EP), precipitable water (PRW), evaporation (LH), and near-surface relative humidity (RHS). The utility of P and EP as model diagnostics was highlighted by Greve et al., (2018) in selecting a subset of CMIP5 models. As global evaporation fields from direct observations and estimated from satellite also contain substantial uncertainty, precipitation minus evaporation is estimated here instead from the vertically integrated divergence of moisture simulated in ERA-Interim fields, which is also arguably the most accurate means of evaluating large-scale patterns and variability (Trenberth and Fasullo, 2013). Precipitation is estimated from the Global Precipitation Climatology Project (GPCP, Adler et al., 2020) Climate Data Record (Adler et al., 2016). The improved version takes advantage of improvements in the gauge records used for calibration and indirect precipitation estimation from longwave radiances provided by NOAA LEO-IR data. For other water cycle fields, output from the European Centre for Medium Range Weather Forecasts (ECMWF) Reanalysis Version 5 (ERA5; Hersbach et al., 2019) is used. ERA5 is the successor to ERA-Interim, increasing the resolution of reported fields, the range of fields assimilated from satellite instruments, and the simulation accuracy as compared against a broad range of observations for various measures. For example, a comparison of ERA5 to satellite data (CERES, GPCP) demonstrates reduced mean state annual and seasonal biases as compared to ERA-Interim (not shown).

Table 1Sorted summary of CMIP models considered in this work, sorted by overall scores.

Download Print Version | Download XLSX

2.1.3 The dynamical realm

Dynamical fields considered include sea level pressure (SLP), wind speed (US), 500 hPa eddy geopotential height (Z500), vertical velocity (W500), and relative humidity (RH500). The use of eddy geopotential rather than total geopotential, which contains significant spatial variance arising from meridional temperature contrasts, is motivated by its ability to resolve our main field of interest – the spatial structure of atmospheric circulations. ERA5, discussed above, is used for estimation of dynamical fields, as such fields are generally not provided from satellite (excepting RH500). Motivating its use – and among its notable improvements relative to earlier reanalyses – is ERA5's improved representation of tropospheric waves and jets that is core to our dynamical evaluation.

https://gmd.copernicus.org/articles/13/3627/2020/gmd-13-3627-2020-f01

Figure 1Mean simulated fields of SWCF in CESM2 from 1995 to 2014 for (a) the annual mean, (b) seasonal contrasts, and (c) regressed against Niño3.4 SST anomalies using July through June averages. Observed CERES EBAF4.1 estimated SWCF for 2000–2018 for (d–f) analogous metrics and (g–i) CESM2–CERES differences are also shown. Stippling indicates regions where CESM2–CERES differences exceed twice the estimated internal spread from CESM1-LE. Hatching indicates regions where differences exceed the same spread plus observational uncertainty (added in quadrature, applied to all panels in each column). Units are watts per square meter (W m−2) except for regressions (right column), where units are watts per square meter per kelvin (Wm-2K-1). Zonal means (right panels) include land (red), ocean (blue), and global (black).

2.2 Generation of variable, realm, timescale, and overall scores

Scores for annual mean, seasonal mean, and ENSO timescale metrics are generated from the area-weighted pattern correlations (Rs) between each simulated variable and the corresponding observational dataset. Weighted averages of these three Rs values are then used to generate a variable score for each field in a given simulation. Arithmetic averages across the relevant variable scores are then used to generate realm scores, and the realm scores for a simulation are arithmetically averaged to generate an overall score. Similarly, timescale scores are generated by averaging Rs values for the relevant timescale across all variables. The inclusion of both realm and timescale scores is motivated in part by the need to interpret the origin of changes in overall scores, which include a large number of Rs values that may otherwise obscure an obvious physical interpretation. Insights gained, for example, include the attribution of much of the overall score improvement across CMIP generations to the fidelity of simulated ENSO patterns.

https://gmd.copernicus.org/articles/13/3627/2020/gmd-13-3627-2020-f02

Figure 2Overall, realm, timescale, and variable scores (ordinate) for historical (20c3m) simulations submitted to the CMIP3 archives (abscissa), sorted by overall score (top row) based on methods employed (see Methods). Simulations and variables are ordered in descending score order from left to right using the overall score and from top to bottom using average variable score.

Download

The use of weights in generating variable scores is motivated by the desire to assist in interpretation of differences in the overall score relative to the influence of internal variability. Using the Community Earth System Version 1 Large Ensemble (CESM1-LE; Kay et al., 2015), weights for ENSO scores are reduced from 1.000 to 0.978 (while for annual and seasonal scores they are 1.000) such that the standard deviation range in overall scores for the 40 members of the CESM1-LE is 0.010. This therefore can be used to interpret generally the approximate contribution of internal variability to intermodel overall scores in analysis of the CMIP archives, suggesting that differences between individual simulations of less than approximately 0.040 (±2σ) are not statistically significant. Where available, multiple-simulation analyses provide an opportunity for further narrowing the uncertainty in statements regarding intermodel fidelity, and as will be seen, overall score ranges within and across the CMIP ensembles generally exceed the obscuring effects of internal variability.

2.3 CMIP Simulations

As the goal of this work is to characterize the evolution of agreement between climate models generally across the CMIP archives and the observations, all available model submissions for which sufficient data are provided are included in the analysis (as summarized in Table 1). A major exception to the data availability requirement relates to near-surface wind speed (US), which was not included as part of the CMIP3 variable list specification. Scores for the dynamical realm in CMIP3 therefore omit US as a scored variable and instead compute the dynamic realm score from the remaining dynamic variable scores. While multiple ensemble members are provided in the CMIP archives for many models and have been assessed, only a single member of each model is incorporated into the analysis here to avoid overweighting the influence of any single mode.

Lastly, in an effort to quantify the leading patterns of bias that differentiate models, a covariance-matrix-based principal component (PC) analysis is used where the array of bias patterns (long × lat × model) is decomposed for its empirical orthogonal functions (EOFs). The EOFs are plotted as regressions against the normalized PC time series and therefore have the same units as the raw fields. Shown are the two leading EOFs and corresponding PC values, sorted by their values and averaged across terciles for each CMIP generation. Included in the PC analysis is an observational estimate (i.e., zero bias) to provide context for model differences. The leading EOFs are found to be both separable and explain significant variance in the bias matrix.

3 Assessing CMIP scores

To illustrate the analysis approach and provide context for the magnitude of biases relative to internal variability and observational uncertainty, Fig. 1 shows both observed and simulated SWCF fields across the timescales considered (Fig. 1a, annual, Fig. 1b, seasonal, and Fig. 1c, ENSO) in the CESM Version 2 submission to CMIP6, CERES estimates (Fig. 1d–f), and their differences (CESM2–CERES, Fig. 1g–i). Significant spatial structure characterizes all fields, with a strong SWCF cooling influence in the mean across much of the globe (Fig. 1a), seasonal contrasts (Fig. 1b) that vary between land and ocean and latitudinal zone, and ENSO teleconnections (Fig. 1c) that extend from the tropical Pacific Ocean to remote ocean basins and the extratropics. While (as will be seen) CESM2 scores among the best available climate models, large model–observation differences nonetheless exist. Regions where model–observation differences are larger than twice the ensemble standard deviation in the CESM1-LE in the annual and seasonal means (stippled) are widespread and remain extensive where the uncertainty range is expanded to incorporate estimated observational uncertainty (added in quadrature, hatched) from Loeb et al. (2018). Of particular note is the fact that it is the large-scale coherent patterns of bias, where model–observational disagreement exceeds uncertainty bounds, that are the primary drivers of pattern correlations used in scoring, rather than synoptic-scale noise.

https://gmd.copernicus.org/articles/13/3627/2020/gmd-13-3627-2020-f03

Figure 3As in Fig. 2 except for historical simulations submitted to the CMIP5 archive.

Download

https://gmd.copernicus.org/articles/13/3627/2020/gmd-13-3627-2020-f04

Figure 4As in Fig. 2 except for historical simulations submitted to the CMIP6 archive.

Download

The color table summary of scores for CMIP3 (mean pattern correlations scaled by 100, Fig. 2) provides a visual summary of simulation performance across the models in the archive (abscissa), including variable, realm, timescale, and overall scores (i.e., aggregate scores, ordinate). Simulations are sorted by overall scores (top row, descending scores toward the right). Realm and timescale scores (rows 2 through 7) also provide broad summaries of model performance. Mean overall scores (69±7, 1σ) are modest generally in CMIP3 and generally uniform across realms. CMIP3 simulations score particularly poorly for ENSO, where scores average to 47, are generally less than 60, and approach 0 in some coarse-grid models. Variable scores are highest for PRW and OLR (which are strongly tied to surface temperature) – and for SLP – and less for other variables, with the lowest scores reported for RS and W500. Spread across models for RS is particularly large relative to other variables. Average variable scores are also poor for SWCF (68), LWCF (71), and P (69), which are among the more important simulated fields according to expert consensus (Burrows et al., 2018).

The color table summary of scores for CMIP5 (Fig. 3) reveals scores that are considerably higher than most CMIP3 simulations, with improvements in the average overall score of (75±5) and most notable improvements on the ENSO timescale, with an average of 57, though with considerable intermodel range (σ=10). A broad increase in scores in the highest-performing models is apparent with numerous variable scores exceeding 85 (orange/red) and several overall scores of 80 or better. As for CMIP3 the highest-scoring variables are PRW, SLP, and OLR, while RHS and W500 are among the lowest-scoring variables. Mean variable scores remain relatively low for SWCF (71), LWCF (75), and P (73).

The color table summary of scores for CMIP6 (Fig. 4) illustrates scores that are considerably higher than both CMIP3 and CMIP5 simulations, with improvements in the average overall score of (79±4) and most continued improvements on the ENSO timescale, though again with considerable intermodel range. A continued increase in scores in the highest-performing models is again apparent, with scores reaching the mid to upper 70s and numerous variable scores exceeding 90 (red). The highest-scoring variables again include PRW, SLP, and OLR, though scores are also high for RH500, one of the more important simulated fields according to expert consensus (Burrows et al., 2018). Scores also increase for SWCF (78), LWCF (80), and P (77).

To highlight connections between variables and aid in identifying the main variables driving variance in aggregate scores across the CMIP archives, correlations amongst scores across all CMIP models are shown in Fig. 5. For overall scores, these include strong connections to P, EP, and OLR, fields strongly connected to atmospheric heating, dynamics, and deep convection and therefore broadly relevant to model performance. Strong connections also exist for SWCF, LWCF, and RH500, consistent with the expert consensus in highlighting these fields as being particularly important (Burrows et al., 2018). An approximately equal correlation exists across realms with the overall score, while for timescales, ENSO exhibits the strongest overall correlation as it contains the greatest intermodel variance and thus explains a greater portion of the overall score variance. Correlations between timescales are weak generally, consistent with the findings of Gleckler et al., (2008) where relationships were also examined between the mean state and interannual variability. Notable as well is that some variables for which scores are high in the mean, such as SLP and PRW, exhibit little correlation with the overall score as the uniformly high scores across models impart relatively little variance to the overall scores.

https://gmd.copernicus.org/articles/13/3627/2020/gmd-13-3627-2020-f05

Figure 5Cross correlations between variable and aggregate scores computed for the all CMIP archives sorted in order of decreasing correlations from left to right and top to bottom.

Download

4 Derived bias patterns for selected variables

The observational estimate of SWCF from CERES is shown in Fig. 6a along with mean bias patterns for CMIP3 (b) and CMIP6 (c). A principal component (PC) analysis of the bias across the broader CMIP archives is also conducted (see Methods) with the leading principal components and their tercile mean values within each CMIP version being shown (d) along with the two leading patterns of bias (Fig. 6e and f). The mean observational field (Fig. 6a) is characterized by negative values in nearly all locations (except over ice) and the strongest cooling influence in the deep tropics, subtropical stratocumulus regions, and midlatitude oceans. Mean bias patterns demonstrate considerable improvement across the CMIP generations, with major reductions in negative biases in the subtropical and tropical oceans. Variance across models is characterized by the degree of tropical–extratropical contrasts in SWCF (EOF1), which explains 24 % of the intermodel variance, and land–ocean contrasts (EOF2), which explain 16 % of the variance. The expression of both patterns of biases is demonstrated to diminish across CMIP generations and terciles in their PC weights (Fig. 6d), ordered sequentially (1–3) with CMIP6 values (dark blue) lying generally closer to observations than CMIP3/5. Improvements are not however necessarily monotonic across the CMIP generations, with improvements and degradations notable in some aspects of the PC1/2 transition from CMIP3 to CMIP5 (i.e., instances in which tercile mean PC values are closer to CERES for CMIP3 than CMIP5).

https://gmd.copernicus.org/articles/13/3627/2020/gmd-13-3627-2020-f06

Figure 6Analysis of the annual mean SWCF bias in the combined historical CMIP3/5/6 archive including the following: (a) the observed estimate from CERES EBAFv4.1; the mean biases in (b) CMIP3 and (c) CMIP6; and (d) the first two PCs of biases and their tercile averages across the CMIP archives and the associated (e) first and (f) second EOFs of biases. All units are watts per square meter (W m−2), except for the PCs, which are unitless. Zonal means (right panels) include land (red), ocean (blue), and global (black).

https://gmd.copernicus.org/articles/13/3627/2020/gmd-13-3627-2020-f07

Figure 7Analysis of the annual mean LWCF bias in the combined historical CMIP3/5/6 archive including the following: (a) the observed estimate from CERES EBAFv4.1; the mean biases in (b) CMIP3 and (c) CMIP6; and (d) the first two PCs of biases and their tercile averages across the CMIP archives and the associated (e) first and (f) second EOFs of biases. All units are watts per square meter (W m−2), except for the PCs, which are unitless. Zonal means (right panels) include land (red), ocean (blue), and global (black).

The observational estimate for LWCF from CERES is shown in Fig. 7a along with mean bias patterns for CMIP3 (b) and CMIP6 (c). A PC analysis of the bias across the CMIP archives is also shown with the leading PC weights and their tercile mean values within each CMIP version being shown (d) along with the two leading patterns of bias (Fig. 7e, f). Observational fields are characterized by a strong heating influence in regions of deep tropical convection and in the extratropical ocean regions in which SWCF is also strong (Fig. 6a), while weak heating is evident in the subtropics and polar regions. Significant changes characterize mean bias patterns between CMIP3 and CMIP6, with positive biases across most ocean regions in CMIP3 and negative biases in many of the same regions in CMIP6. On average however, the magnitudes of biases are reduced across CMIP generations. This is evident for example in the PC analysis of bias (Fig. 7d), where CMIP6 values lie closer generally to CERES than for CMIP3 or CMIP5. The leading mode (EOF1, Fig. 7e) exhibits strong weightings over the warm pool, is negatively correlated with both the mean pattern and bias, and explains 36 % of the intermodel variance. In contrast, EOF2 exhibits a strong tropical–extratropical contrast, has little correlation to the mean pattern or bias, and explains only 13 % of the variance. The PC1/2 tercile weights for these modes show a considerable reduction in EOF1 spread, smaller mean tercile biases generally, and improved agreement across model terciles from CMIP3 to CMIP6, though as with SWCF, the improvement is not monotonic nor uniform across all terciles and PCs.

https://gmd.copernicus.org/articles/13/3627/2020/gmd-13-3627-2020-f08

Figure 8Analysis of the annual mean precipitation bias in the combined historical CMIP3/5/6 archive including the following: (a) the observed estimate from GPCP CDR; the mean biases in (b) CMIP3 and (c) CMIP6; and (d) the first two PCs of biases and their tercile averages across the CMIP archives and the associated (e) first and (f) second EOFs of biases. All units are millimeters per day (mm d−1), except for the PCs, which are unitless. Zonal means (right panels) include land (red), ocean (blue), and global (black).

The observational estimate for precipitation from GPCP is shown in Fig. 8a along with mean bias patterns for CMIP3 (Fig. 8b) and CMIP6 (Fig. 8c). The PC analysis of the bias across the CMIP archives is also shown with the leading PC tercile mean values for each CMIP version being shown (Fig. 8d) along with the two leading patterns of bias (Fig. 8e, f). The annual mean pattern resolves key climate system features, including strong precipitation in the Inter-Tropical Convergence Zone (ITCZ) and arid conditions in the subtropics and at high latitudes. Biases are large in both CMIP3 and CMIP6 on average and are characterized generally by excessive subtropical precipitation and deficient precipitation in the Pacific Ocean ITCZ, in South America, and at high latitudes. Earlier work has generally characterized model bias in terms of its double-ITCZ structure (Oueslati and Bellon, 2015), though systematic bias is also apparent beyond the tropical Pacific Ocean. In addition, the PC decomposition of CMIP precipitation biases (Fig. 8d–f) suggests that the bias is comprised to two orthogonal leading patterns that together explain 15 % and 11 % of the variance across models, respectively. A separable unique leading pattern is therefore not evident. Rather, the leading pattern (Fig. 8e) is characterized by weakness in precipitation across the equatorial oceans, with elevated rates in the Maritime continent and in the Pacific Ocean near 15 N and 15 S. The second pattern (Fig. 8f) is characterized by loadings over Africa and South America and on the southern fringe of the observed climatological Pacific ITCZ (Fig. 8a), with negative loadings in the subtropical ocean basins. Based on mean PC tercile values, slight improvement across CMIP generations is evident, as tercile values lie closer to observations for all terciles of PC1/2 in CMIP6 versus CMIP3, with the exception of the first tercile of PC1, where CMIP3 lies close to GPCP.

https://gmd.copernicus.org/articles/13/3627/2020/gmd-13-3627-2020-f09

Figure 9Analysis of the annual mean RH500 bias in the combined historical CMIP3/5/6 archive including the following: (a) the observed estimate from ERA5; the mean bias in (b) CMIP3 and (c) CMIP6; and (d) the first two PCs of biases and their tercile averages across the CMIP archives and the associated (e) first and (f) second EOFs of biases. All units are percent (%), except for the PCs, which are unitless. Zonal means (right panels) include land (red), ocean (blue), and global (black).

The observational estimate for RH500 from ERA5 is shown in Fig. 9a along with mean bias patterns for CMIP3 (b) and CMIP6 (c). A principal component analysis of the bias across the CMIP archives is also shown with the leading principal components and their tercile mean values within each CMIP version being shown (Fig. 9d) along with the two leading patterns of bias (Fig. 9e and f). The observed RH500 field is characterized by positive humidity biases in regions of frequent deep convection (i.e., Maritime continent, Amazon) and at high latitudes and very dry conditions in the subtropics, with values generally below 30 % across the subtropics, features that were poorly resolved in CMIP3 (e.g., Fasullo and Trenberth 2012). The CMIP3 mean bias field is negatively correlated with the mean state, with patterns that lack sufficient spatial contrast, are too moist in the subtropics, and are too dry in Africa, the Maritime continent, the Amazon, and at high latitudes. The magnitude of mean RH500 biases in CMIP6 are substantially smaller (roughly 50 %) than CMIP3, though they share a similar overall pattern reflecting weakness in spatial contrasts. The PC analysis of bias reveals a leading pattern that explains 50 % of the intermodal variance and is negatively correlated with observations (−0.44). The second leading pattern (Fig. 9f) explains considerably less variance (15 %) and exhibits a zonally uniform structure characterized by tropical–extratropical contrast. The weights for PC1/2 reveal systematic bias in PC1 across models (all lie to the right of ERA5) and considerable improvement across CMIP generations as CMIP6 weights lie significantly closer to ERA5 than CMIP3 weights for all terciles (1–3). Small improvements are also evident in terciles 1 and 2 of PC2, though this comprises a small fraction of variance in overall CMIP bias.

https://gmd.copernicus.org/articles/13/3627/2020/gmd-13-3627-2020-f10

Figure 10Evolution of the distribution of aggregate and selected variable scores across the CMIP archives and the CESM1-LE.

Download

In the effort to summarize the evolution of the full distributions of scores across the CMIP archives, whisker plots encompassing the median, interquartile, and 10–90 percentile ranges are shown for various aggregate metrics and key fields in Fig. 10. Also shown are the equivalent ranges for scores computed from the CESM1-LE to provide an estimate of the influence of internal variability for each distribution. A steady improvement in the overall scores is evident across CMIP versions, a progression that is also evident across realm scores and particularly for the poorest scoring models in the dynamics realm. Scores for annual and seasonal timescales are generally high across archives, though internal variability is also small and is substantially less than the median improvements across the archives. The range of scores for ENSO is significantly greater than other timescales, as is the range of internal variability, and substantial improvements have been realized for the lowest-scoring models across successive CMIP generations. Noteworthy are the substantial improvements in SWCF, LWCF, and P, with the best CMIP3 simulations scoring near the median value for CMIP6 and improvements in median values from CMIP3 to CMIP6 exceeding uncertainty arising from internal variability. Scores for RH500 have also improved, although the spread within the CMIP3 archives is substantial, and uncertainty arising from internal variability is somewhat greater than for other variables. RH500 scores in CMIP6 are generally higher than for cloud forcing and P. For SLP, median scores are uniformly high across the CMIP generations, with small but steady improvement in median and interquartile scores, with the main exception of high scores being the low-scoring 0 %–25 % range of CMIP3 simulations.

5 Discussion

An objective model evaluation approach has been developed that uses feedback-relevant fields and takes advantage of recent expert elicitations of the climate modeling community and advances in satellite and reanalysis datasets. In its application to the CMIP archives, the analysis is shown to provide an objective means for computing model scores across variables, realms, and timescales. Visual summaries of model performance across the CMIP archives are also generated, which readily allow for the survey of a broad suite of climate performance scores. As there is unlikely to be a single model best suited to all applications (Gleckler et al., 2008; Knutti et al., 2010, 2017), in providing online access to model scores and the fields used to compute them, the results herein are intended to aid the community in informing model ensemble optimization for targeted applications.

Based on the pattern correlation approach adopted, a number of statements can be made regarding the overall performance of climate models across CMIP generations. Noteworthy is that, as informed by analysis of the CESM1-LE and consistent with the design of the approach used, these statements are robust to the obscuring influence of internal climate variability. In general, computed scores have increased steadily across CMIP generations, with improvements exceeding the range of internal variability. Associated with these improvements, the leading patterns of bias across models are shown to have been reduced. Improvements are large and particularly noteworthy for ENSO teleconnection patterns, as the poorest scoring models in each CMIP generation have improved substantially. In part this may be due to the elimination of very low resolution models in CMIP5/6, though improvements in model physics is also likely to play a role. The overall range of model performance within CMIP versions has also decreased in conjunction with increases in median scores, as improvement in the worst models has generally outpaced that of the median. Reductions in systematic patterns of bias (e.g., Figs. 6–9) across the CMIP archives have been pronounced for fields deemed in expert solicitations to have particular importance, including SWCF, LWCF, and RH500.

Also relevant for climate feedbacks, variable scores for SWCF, LWCF, RH500, and precipitation have increased steadily across the CMIP generations (e.g., Fig. 10), with magnitudes exceeding the uncertainty associated with internal variability. Scores are particularly high for CMIP6 models for which high climate sensitivities have been reported, including CESM2, SAM0-UNICON, GFDL-CM4, CNRM-CM6-1, E3SM, and EC-Earth3-Veg (though exceptions also exist such as in the case of MIRCO6). These findings therefore echo the concerns voiced in Gettelman et al. (2019): “What scares us is not that the CESM2 ECS is wrong (all models are wrong; Box, 1976) but that it might be right.”. The fields provided by the Climate Model Assessment Tool (CMAT) allow for an expedited analysis of the sources of these improvements, such as for example the simulation of supercooled liquid clouds (e.g., Kay et al., 2016). Further work examining the ties between metrics of performance in simulating the present-day climate, such as those provided here, and longer-term climate model behavior is warranted to bolster confidence in model projections of climate change.

Code and data availability

Data used in this study are available freely from the Earth System Grid at https://www.earthsystemgrid.org (last access: 18 August 2020).

NetCDF output for the fields generated herein is freely available at http://webext.cgd.ucar.edu/Multi-Case/CMAT/index.html (last access: 18 August 2020). The software used to develop this work is available under a free software license at Zenodo: https://doi.org/10.5281/zenodo.3922308 (Fasullo, 2020).

Author contributions

JF designed the analysis routine, performed the model analysis, developed the analysis, obtained the data from the ESG and created all graphics. JF composed the paper.

Competing interests

The author declares that there is no conflict of interest.

Acknowledgements

This material is based upon work supported by the National Center for Atmospheric Research, which is a major facility sponsored by the National Science Foundation under cooperative agreement no. 1852977. The efforts of John T. Fasullo in this work were supported by NASA award 80NSSC17K0565, by NSF award #AGS-1419571, and by the Regional and Global Model Analysis (RGMA) component of the Earth and Environmental System Modeling Program of the U.S. Department of Energy's Office of Biological & Environmental Research (BER) via National Science Foundation IA 1844590.

Financial support

This research has been supported by the National Science Foundation (grant nos. 1852977, AGS-1419571) and the NASA (grant no. 80NSSC17K0565).

Review statement

This paper was edited by Juan Antonio Añel and reviewed by three anonymous referees.

References

Adler, R., Sapiano, M., Huffman, G., Bolvin, D., Gu, G., Wang, J., and Schneider, U.: The new version 2.3 of the Global Precipitation Climatology Project (GPCP) monthly analysis product, University of Maryland, April, pp. 1072–1084, 2016. 

Adler, R. F., Gu, G., Huffman, G. J., Sapiano, M. R., and Wang, J. J.: GPCP and the Global Characteristics of Precipitation, in: Satellite Precipitation Measurement, pp. 677–697, Springer, Cham, 2020. 

Baker, N. C. and Taylor, P. C.: A framework for evaluating climate model performance metrics, J. Climate, 29, 1773–1782, https://doi.org/10.1175/JCLI-D-15-0114.1, 2016. 

Borovikov, A., Cullather, R., Kovach, R., Marshak, J., Vernieres, G., Vikhliaev, Y., and Li, Z.: GEOS-5 seasonal forecast system, Clim. Dynam., 53, 7335–7361, https://doi.org/10.1007/s00382-017-3835-2, 2019. 

Box, G. E. P.: Science and statistics, J. Am. Stat. Assoc., 71, 791–799, https://doi.org/10.1080/01621459.1976.10480949, 1976. 

Braverman, A., Chatterjee, S., Heyman, M., and Cressie, N.: Probabilistic evaluation of competing climate models, Adv. Stat. Clim. Meteorol. Oceanogr., 3, 93–105, https://doi.org/10.5194/ascmo-3-93-2017, 2017. 

Burrows, S. M., Dasgupta, A., Reehl, S., Bramer, L., Ma, P. L., Rasch, P. J., and Qian, Y.: Characterizing the relative importance assigned to physical variables by climate scientists when assessing atmospheric climate model fidelity, Adv. Atmos. Sci., 35, 1101–1113, https://doi.org/10.1007/s00376-018-7300-x, 2018. 

Cheng, L., Trenberth, K. E., Fasullo, J. T., Mayer, M., Balmaseda, M., and Zhu, J.: Evolution of ocean heat content related to ENSO, J. Clim., 32, 3529–3556, https://doi.org/10.1175/JCLI-D-18-0607.1, 2019. 

Dee, D. P., Uppala, S. M., Simmons, A. J., Berrisford, P., Poli, P., Kobayashi, S., and Bechtold, P.: The ERA-Interim reanalysis: Configuration and performance of the data assimilation system, Q. J. Roy. Meteorol. Soc., 137, 553–597, https://doi.org/10.1002/qj.828, 2011. 

Eyring, V., Bony, S., Meehl, G. A., Senior, C. A., Stevens, B., Stouffer, R. J., and Taylor, K. E.: Overview of the Coupled Model Intercomparison Project Phase 6 (CMIP6) experimental design and organization, Geosci. Model Dev., 9, 1937–1958, https://doi.org/10.5194/gmd-9-1937-2016, 2016. 

Eyring, V., Bock, L., Lauer, A., Righi, M., Schlund, M., Andela, B., Arnone, E., Bellprat, O., Brötz, B., Caron, L.-P., Carvalhais, N., Cionni, I., Cortesi, N., Crezee, B., Davin, E. L., Davini, P., Debeire, K., de Mora, L., Deser, C., Docquier, D., Earnshaw, P., Ehbrecht, C., Gier, B. K., Gonzalez-Reviriego, N., Goodman, P., Hagemann, S., Hardiman, S., Hassler, B., Hunter, A., Kadow, C., Kindermann, S., Koirala, S., Koldunov, N., Lejeune, Q., Lembo, V., Lovato, T., Lucarini, V., Massonnet, F., Müller, B., Pandde, A., Pérez-Zanón, N., Phillips, A., Predoi, V., Russell, J., Sellar, A., Serva, F., Stacke, T., Swaminathan, R., Torralba, V., Vegas-Regidor, J., von Hardenberg, J., Weigel, K., and Zimmermann, K.: Earth System Model Evaluation Tool (ESMValTool) v2.0 – an extended set of large-scale diagnostics for quasi-operational and comprehensive evaluation of Earth system models in CMIP, Geosci. Model Dev., 13, 3383–3438, https://doi.org/10.5194/gmd-13-3383-2020, 2020. 

Fasullo, J.: Climate Model Analysis Tool (Version 1), Zenodo, https://doi.org/10.5281/zenodo.3922308, 2020. 

Fasullo, J. T. and Trenberth, K. E.: The annual cycle of the energy budget. Part I: Global mean and land–ocean exchanges. J. Climate, 21, 2297–2312, https://doi.org/10.1175/2007JCLI1935.1, 2008. 

Fasullo, J. T. and Trenberth, K. E.: A less cloudy future: The role of subtropical subsidence in climate sensitivity, Science, 338, 792–794, https://doi.org/10.1126/science.1227465, 2012. 

Gettelman, A., Hannay, C., Bacmeister, J. T., Neale, R., Pendergrass, A. G., Danabasoglu, G., and Mills, M. J.: High climate sensitivity in the Community Earth System Model Version 2 (CESM2), Geophys. Res. Lett., 46, 8329–8337, https://doi.org/10.1029/2019GL083978, 2019. 

Ghil, M.: A mathematical theory of climate sensitivity or, How to deal with both anthropogenic forcing and natural variability? In: Climate Change: Multidecadal and Beyond (vol. 6, pp. 31–52), edited by: Chang, C.-P., Ghil, M., Latif, M., and Wallace, J., World Scientific Publishing Co., Singapore, 2015. 

Gibbs, A. L. and Su, F. E.: On Choosing and Bounding Probability Metrics, Int. Stat. Rev., 70, 419–435, https://doi.org/10.1111/j.1751-5823.2002.tb00178.x, 2002. 

Gleckler, P. J., Taylor, K. E., and Doutriaux, C.: Performance metrics for climate models, J. Geophys. Res.-Atmos., 113, 1–20, https://doi.org/10.1029/2007JD008972, 2008. 

Caldwell, P. M., Van Roekel, L. P., Petersen, M. R., Tang, Q., Wolfe, J. D., Abeshu, G., Anantharaj, V., Asay‐Davis, X. S., Bader, D. C., Baldwin, S. A., Bisht, G., Bogenschutz, P. A., Branstetter, M., Brunke, M. A., Brus, S. R., Burrows, S. M., Cameron‐Smith, P. J., Donahue, A. S., Deakin, M., Easter, R. C., Evans, K. J., Feng, Y., Flanner, M., Foucar, J. G., Fyke, J. G., Griffin, B. M., Hannay, C., Harrop, B. E., Hoffman, M. J., Hunke, E. C., Jacob, R. L., Jacobsen, D. W., Jeffery, N., Jones, P. W., Keen, N. D., Klein, S. A., Larson, V. E., Leung, L. R., Li, H.‐Y., Lin, W., Lipscomb, W. H., Ma, P.‐L., Mahajan, S., Maltrud, M. E., Mametjanov, A., McClean, J. L., McCoy, R. B., Neale, R. B., Price, S. F., Qian, Y., Rasch, P. J., Reeves Eyre, J. E. J., Riley, W. J., Ringler, T. D., Roberts, A. F., Roesler, E. L., Salinger, A. G., Shaheen, Z., Shi, X., Singh, B., Tang, J., Taylor, M. A., Thornton, P. E., Turner, A. K., Veneziani, M., Wan, H., Wang, H., Wang, S., Williams, D. N., Wolfram, P. J., Worley, P. H., Xie, S., Yang, Y., Yoon, J.‐H., Zelinka, M. D., Zender, C. S., Zeng, X., Zhang, C., Zhang, K., Zhang, Y., Zheng, X., Zhou, T., and Zhu, Q.: The DOE E3SM coupled model version 1: Overview and evaluation at standard resolution, J. Adv. Model. Earth Syst. 11, 2089–2129, https://doi.org/10.1029/2018MS001603, 2019. 

Greve, P., Gudmundsson, L., and Seneviratne, S. I.: Regional scaling of annual mean precipitation and water availability with global temperature change, Earth Syst. Dynam., 9, 227–240, https://doi.org/10.5194/esd-9-227-2018, 2018. 

Hersbach, H., Bell, B., Berrisford, P., Horányi, A., Sabater, J. M., Nicolas, J., Radu, R., Schepers, D., Simmons, A., Soci, C., and Dee, D.: Global reanalysis: goodbye ERA-Interim, hello ERA5, ECMWF, https://doi.org/10.21957/vf291hehd7, 2019. 

Hourdin, F., Mauritsen, T., Gettelman, A., Golaz, J.-C., Balaji, V., Duan, Q., Folini, D., Ji, D., Klocke, D., Qian, Y., Rauser, F., Rio, C., Tomassini, L., Watanabe, M., and Williamson, D.: The Art and Science of Climate Model Tuning, B. Am. Meteorol. Soc., 98, 589–602, https://doi.org/10.1175/BAMS-D-15-00135.1, 2017. 

Hunt, B. G. and Manabe, S.: Experiments with a stratospheric general circulation model: II. Large-scale diffusion of tracers in the stratosphere, Mon. Weather Rev., 96, 503–539, https://doi.org/10.1175/1520-0493(1968)096<0503:EWASGC>2.0.CO;2, 2009. 

Kay, J. E., Deser, C., Phillips, A., Mai, A., Hannay, C., Strand G., and Holland, M.: The Community Earth System Model (CESM) large ensemble project: A community resource for studying climate change in the presence of internal climate variability, B. Am. Meteorol. Soc., 96, 1333–1349, https://doi.org/10.1175/BAMS-D-13-00255.1, 2015. 

Kay, J. E., Wall, C., Yettella, V., Medeiros, B., Hannay, C., Caldwell, P., and Bitz, C.: Global climate impacts of fixing the Southern Ocean shortwave radiation bias in the Community Earth System Model (CESM), J. Clim., 29, 4617–4636, https://doi.org/10.1175/JCLI-D-15-0358.1, 2016. 

Knutti, R., Abramowitz, G., Collins, M., Eyring, V., Gleckler, P. J., Hewitson, B., and Mearns, L.: Good Practice Guidance Paper on Assessing and Combining Multi Model Climate Projections, in: Meeting Report of the Intergovernmental Panel on Climate Change Expert Meeting on Assessing and Combining Multi Model Climate Projections, edited by: Stocker, T. F., Qin, D., Plattner, G.-K., Tignor, M., Midgley, P. M., IPCC Working Group I Technical Support Unit, University of Bern, Bern, Switzerland, 2010. 

Knutti, R., Sedláček, J., Sanderson, B. M., Lorenz, R., Fischer, E. M., and Eyring, V.: A climate model projection weighting scheme accounting for performance and interdependence, Geophys. Res. Lett., 44, 1909–1918, https://doi.org/10.1002/2016GL072012, 2017. 

Lembo, V., Lunkeit, F., and Lucarini, V.: TheDiaTo (v1.0) – a new diagnostic tool for water, energy and entropy budgets in climate models, Geosci. Model Dev., 12, 3805–3834, https://doi.org/10.5194/gmd-12-3805-2019, 2019. 

Loeb, N. G., Doelling, D. R., Wang, H., Su, W., Nguyen, C., Corbett, J. G., and Kato, S.: Clouds and the earth's radiant energy system (CERES) energy balanced and filled (EBAF) top-of-atmosphere (TOA) edition-4.0 data product, J. Clim., 31, 895–918, https://doi.org/10.1175/JCLI-D-17-0208.1, 2018. 

Lorenz, R., Herger, N., Sedláček, J., Eyring, V., Fischer, E. M., and Knutti, R.: Prospects and caveats of weighting climate models for summer maximum temperature projections over North America, J. Geophys. Res.-Atmos., 123, 4509–4526, https://doi.org/10.1029/2017JD027992, 2018. 

Manabe, S., Bryan, K., and Spelman, M. J.: A global ocean-atmosphere climate model. Part I. The atmospheric circulation, J. Phys. Ocean., 5, 3–29, https://doi.org/10.1175/1520-0485(1975)005<0003:AGOACM>2.0.CO;2, 1975. 

Meehl, G. A., Covey, C., Delworth, T., Latif, M., McAvaney, B., Mitchell, J. F. B., Stouffer, R. J., and Taylor, K. E.: THE WCRP CMIP3 Multimodel Dataset: A New Era in Climate Change Research, B. Am. Meteorol. Soc., 88, 1383–1394, https://doi.org/10.1175/BAMS-88-9-1383, 2007. 

Neubauer, D., Ferrachat, S., Siegenthaler-Le Drian, C., Stier, P., Partridge, D. G., Tegen, I., Bey, I., Stanelle, T., Kokkola, H., and Lohmann, U.: The global aerosol–climate model ECHAM6.3–HAM2.3 – Part 2: Cloud evaluation, aerosol radiative forcing, and climate sensitivity, Geosci. Model Dev., 12, 3609–3639, https://doi.org/10.5194/gmd-12-3609-2019, 2019.  

Oueslati, B. and Bellon, G.: The double ITCZ bias in CMIP5 models: interaction between SST, large-scale circulation and precipitation, Clim. Dynam., 44, 585–607, https://doi.org/10.1007/s00382-015-2468-6, 2015. 

Pierce, D. W., Barnett, T. P., Santer, B. D., and Gleckler, P. J.: Selecting global climate models for regional climate change studies, P. Natl. Acad. Sci. USA, 106, 8441–8446, https://doi.org/10.1073/pnas.0900094106, 2009. 

Pincus, R., Batstone, C. P., Hofmann, R. J. P., Taylor, K. E., and Glecker, P. J.: Evaluating the present-day simulation of clouds, precipitation, and radiation in climate models, J. Geophys. Res.-Atmos., 113, D14209, https://doi.org/10.1029/2007JD009334, 2008. 

Schmidt, G. A., Bader, D., Donner, L. J., Elsaesser, G. S., Golaz, J.-C., Hannay, C., Molod, A., Neale, R. B., and Saha, S.: Practice and philosophy of climate model tuning across six US modeling centers, Geosci. Model Dev., 10, 3207–3223, https://doi.org/10.5194/gmd-10-3207-2017, 2017. 

Taylor, K. E., Stouffer, R. J., and Meehl, G. A.: An overview of CMIP5 and the experiment design, B. Am. Meteorol. Soc., 93, 485–498, https://doi.org/10.1175/BAMS-D-11-00094.1, 2012. 

Trenberth, K. E. and Fasullo, J. T.: Regional energy and water cycles: Transports from ocean to land, J. Clim., 26, 7837–7851, https://doi.org/10.1175/JCLI-D-13-00008.1, 2013. 

Trenberth, K. E. and Fasullo, J. T.: Atlantic meridional heat transports computed from balancing Earth's energy locally, Geol. Res. Lett., 44, 1919–1927, https://doi.org/10.1002/2016GL072475, 2017. 

Download
Short summary
The fidelity of climate model simulations included in the WCRP Coupled Model Intercomparison Project Versions 3 through 6 is evaluated using best estimates of fields considered by the community to be critical for climate change projections. The analysis benchmarks patterns of the mean state and variability (seasonal/interannual) both within and across model generations, highlighting progress and quantifying persisting biases across models.