the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
An extensible perturbed parameter ensemble for the Community Atmosphere Model version 6
Trude Eidhammer
Andrew Gettelman
Katherine Thayer-Calder
Duncan Watson-Parris
Gregory Elsaesser
Hugh Morrison
Marcus van Lier-Walqui
Ci Song
Daniel McCoy
This paper documents the methodology and preliminary results from a perturbed parameter ensemble (PPE) technique, where multiple parameters are varied simultaneously and the parameter values are determined with Latin hypercube sampling. This is done with the Community Atmosphere Model version 6 (CAM6), the atmospheric component of the Community Earth System Model version 2 (CESM2). We apply the PPE method to CESM2–CAM6 to understand climate sensitivity to atmospheric physics parameters. The initial simulations vary 45 parameters in the microphysics, convection, turbulence and aerosol schemes with 263 ensemble members. These atmospheric parameters are typically the most uncertain in many climate models. Control simulations and targeted simulations to understand climate forcing due to aerosols and fast climate feedbacks are analyzed. The use of various emulators is explored in the multi-dimensional space mapping input parameters to output metrics. Parameter impacts on various model outputs, such as radiation, cloud and aerosol properties, are evaluated. Machine learning is also used to probe optimal parameter values against observations. Our findings show that PPE is a valuable tool for climate uncertainty analysis. Furthermore, by varying many parameters simultaneously, we find that many different combinations of parameter values can produce results consistent with observations, and thus careful analysis of tuning is important. The CESM2–CAM6 PPE is publicly available and extensible to other configurations to address questions of other model processes in the atmosphere and other model components (e.g., coupling to the land surface).
- Article
(7139 KB) - Full-text XML
- BibTeX
- EndNote
General circulation models (GCMs) have numerous and long-standing biases due in part to uncertain representations of the physical processes (e.g., Trenberth and Fasullo, 2010). This is especially true for processes that occur at subgrid scales, such as microphysics, turbulence, convection and aerosol processes. Because these processes are not resolved, their effects on the grid-scale model state variables are represented via parameterizations rather than the process equations being explicitly solved at the natural scale of the phenomena being represented. For example, the evolution of a single cloud drop in a turbulent flow over a small domain can be simulated explicitly, but the evolution of a cloud drop population cannot be directly simulated directly due to the sheer number of drops within each grid volume of typical atmospheric models (with grid spacing of tens of meters to tens of kilometers). Moreover, for many processes, including cloud and aerosol microphysics, even at the natural scale of the phenomenon (e.g., scale of an individual drop) there are uncertainties in the underlying physical processes. That is to say that for many processes there are no governing equations at any scale. For example, for cloud microphysics there are fundamental uncertainties in how drops collide and either bounce, coalesce or breakup. Most ice microphysical processes, including nucleation, diffusional growth, riming and aggregation, remain highly uncertain even at the scale of individual particles (e.g., Morrison et al., 2020).
Parameterizations typically include parameters whose values are constrained by theory, high-resolution process models and observations. To varying degrees, these parameter values are uncertain because of both uncertainty in how to best represent the impact of subgrid-scale processes at the grid scale and fundamental uncertainty at the process scale. In climate models, parameter values are adjusted within the bounds of uncertainty to produce realistic output relative to observations. However, this process, usually referred to as “tuning”, faces several challenges (Hourdin et al., 2016). For example, since climate GCMs comprise several different physics packages, finding the best parameter values in one physics package could impact the others and produce out-of-balance results. As a consequence, there may be a dependence on the sequence in which the physics packages are tuned. As part of this process, it is important to understand how uncertainty in parameter values translates to uncertainty in simulated climate. Some parameters are more uncertain than others but may have a relatively small or large impact on simulated climate across the range of this uncertainty.
Tuning, and the associated investigation of parameter uncertainties, can be done in several different ways. Each method has an associated computational cost, which is usually a consequence of how many simulations are performed. Traditionally, sensitivity to parameters is analyzed using a “one at a time” (OAT) method (Schmidt et al., 2017). When performed as part of the model tuning process, this can represent an optimized random walk approaching the minimization of an informal cost function (errors against a sum of observations). OAT methods do not account for nonlinear relationships between different parameters and resulting outputs are generally inefficient. Furthermore, to perform simulations over the entire parameter space with many variable parameters, a large number of simulations are required. For example, in the current study we perturb 45 different parameters, which would require a minimum of 3.5 × 1013 (245) simulations using OAT methods if each parameter was tested with only two values in all combinations. The number of simulations needed increases exponentially if each parameter were perturbed with additional values; i.e., the number of required simulations is MN for OAT methods, where N is the number of parameters and M is the number of values tested for each parameter.
Over the last several years, more objective and efficient methods have been developed to perturb multiple moist physics and aerosol parameters simultaneously (Lee et al., 2011; Qian et al., 2015). These methods have been used to optimize models in an automated way (Jackson et al., 2008; Wagman and Jackson, 2018; Regayre et al., 2018; Peatier et al., 2022) and to understand model uncertainty (Posselt and Vukicevic, 2010; van Lier-Walqui et al., 2012; Regayre et al., 2014; Qian et al., 2015; Lee et al., 2016; Qian et al., 2018; Watson-Parris et al., 2020; Duffy et al., 2024; Song et al., 2024; Gettelman et al., 2024). They provide a more robust platform for uncertainty quantification and objective improvement of climate models, ranging from parameter tuning to understanding structural deficiencies of models, for example, when no combination of parameters converges to observations. Comprehensive sets of perturbed parameter values can be used for the development of sophisticated fast model emulators to help with model tuning or even to advance process level understanding and guide selection of key additional data for constraining models (Regayre et al., 2018).
The goal of this paper is to document the methodology for creating a large perturbed parameter ensemble (PPE) with the Community Atmosphere Model version 6 (CAM6; Gettelman et al., 2019), the atmospheric component of the Community Earth System Model version 2 (CESM2; Danabasoglu et al., 2020). We will also present early results on PPE spread for certain outputs, key parameter sensitivities of the model and preliminary results of model emulation. The data, scripts and code for reproducing and extending the PPE are now available to the community. Section 2 contains a description of the methodology used to create the PPE, including parameters and methods. Section 3 describes the method used for the modeling and the emulators. Section 4 describes key initial results of the PPE, emulators and simple tuning, and Sect. 5 provides a summary and conclusions.
To generate the PPE, we created a set of variable parameters using Latin hypercube sampling (McKay et al., 2000). With this technique, random values are created within a determined range. Ranges of the possible values are divided into a number of bins equal to the number of samples. Each parameter is assigned a value within a random bin, and no parameters from subsequent samples can have a value from previously sampled bins. In this way, the parameter sets for all samples cover the entire parameter range for each parameter and have marginal distributions that are uniformly distributed. Figure 1a shows an example of how three different parameters are sampled in relation to each other. The color and size of the symbols represent the distance from the center of origin in the plot to illustrate the depth of the plot. Note that the points in Fig. 1a are generally uniformly distributed in the 3-D space and have uniform marginal distributions in each dimension (Fig. 1b), which is a key aspect of Latin hypercube sampling.
Using this sampling, we initially created 250 different sets of parameter values in addition to the default CESM2–CAM6 setup (total of 251 sets). The ratio of number of ensembles to numbers of parameters is ∼ 5.5, close to the ratio of 6 used in Regayre et al. (2023). After preliminary analysis of the initial simulations we decided to extend the range for one of the parameters (micro_mg_max_nicons). The method we employed is general for any parameters with Latin hypercube sampling. A relative Euclidean distance metric (d) was created. For each individual ensemble m, we calculate the average distance of each parameter i in ensemble m to parameter i in the other ensembles j. Then dm is the sum of all Euclidean distances in ensemble m divided by number of parameters (pa) and ensembles (en):
The relative Euclidean distances for the original 251 ensemble members are shown in Fig. 2 (250 perturbation cases plus 1 default case).
We then generated 7500 new parameter sets. Out of these 7500 sets, we picked 12 sets where the single parameter value of micro_mg_max_nicons was within the new range and had the largest relative Euclidean distance value (equal to or greater than the average Euclidean distance between each of the original 251 parameter samples) to make a total of 263 PPE sets. The reason for choosing the sets with the largest relative Euclidean distance is to avoid the problem of close-proximity points. The relative Euclidean distance of all the 7500 sets that had a parameter value within the new range and a relative Euclidean distance greater than the average distance (0.16) was met for only 32 of the 7500 sets.
Using this approach, we then archived the 262 parameter sets plus the default case in a single file with metadata. Every parameter was chosen to be run-time configurable (not hard-wired in code). A script for running CESM2–CAM6 was developed which sets up a model simulation, then copies (“clones”) the configuration to a new name and substitutes a parameter set from the file. This method enables the reproduction and extension of the PPE from a single file and script. CESM2 and CAM6 can be run in many different configurations (standard atmosphere–ocean for CAM6, fully coupled CESM2, aquaplanet, single column, nudged mode, etc.). Archiving the parameter sets and the automated run script allows any CESM configuration to be run with the same parameter sets for different types of analysis or different diagnostic output.
Here we first describe the CESM2–CAM6 model and then the simulations conducted, parameters varied and finally the emulators used on the model output.
3.1 Model description
In this study, we use CESM2 (Danabasoglu et al., 2020), which contains the atmosphere model CAM6. CAM6 uses a four-mode version of the Liu et al. (2012) Modal Aerosol Model (MAM4) with modifications to include stratospheric sulfur (Mills et al., 2016). This version has an extra mode for primary carbon and has a better representation of black carbon and sulfate evolution. Cloud microphysics in CAM6 uses version 2 of the Morrison and Gettelman (2008) scheme, described by Gettelman and Morrison (2015) and Gettelman et al. (2015). CAM6 replaces the CAM5 shallow convection, planetary boundary layer and cloud macrophysics schemes with a new unified turbulence scheme, the Cloud Layers Unified by Binormals (CLUBB), originally developed by Golaz et al. (2002) and integrated in CAM by Bogenschutz et al. (2013). CAM6 also features a new mixed-phase ice nucleation scheme developed by Hoose et al. (2010). Deep convection is represented by the Zhang and McFarlane (1995) scheme. These CAM6 parameterizations have been implemented in CESM2 as described in Bogenschutz et al. (2018).
3.2 Simulations
We conducted three sets of simulation ensembles using the parameter samples. The first set uses near-present-day cyclic boundary conditions for the year 2000. The greenhouse gases and atmospheric oxidants are average values for the 1995–2005 period. The average monthly sea surface temperatures (SSTs) for 1995–2010 are used. The emission of aerosols and precursors is also set to 1995–2005 in the present day (PD) simulation. The second set of simulations is the preindustrial (PI) configuration. This uses the same setup as PD, but the aerosol emission is estimated for the year 1850. In the third set of simulations the PD configuration is used again, but the SST is uniformly increased by 4 K (SST4K). All simulations use a resolution of 0.9° latitude × 1.1° longitude with 32 levels in the vertical up to 10 hPa. By performing these three sets of simulations with the same parameter sets, not only can we evaluate the output spread by perturbing parameters, but we can also evaluate the cloud feedback (difference between PD and SST4K) and aerosol forcing (difference between PD and PI). Here the aerosol forcing is the aerosol effective forcing after adjustments of atmospheric temperature and humidity. We tested two different run lengths (3 and 5 years) and found that we could reproduce (emulate) a given two-dimensional (2-D) field with similar root mean square error (RMSE) using 3 or 5 years of simulation. All simulations presented herein are 3 years long.
Model output is archived monthly and daily for select fields. Output is available at https://doi.org/10.26024/bzne-yf09 (Eidhammer et al., 2022). Also available is a Python script to create the parameter file and scripts to submit the PPE simulations.
3.3 Parameters
All three simulation sets are run with 263 different ensemble members corresponding to the sets of 45 perturbed parameters plus the default parameter set as described in Sect. 2. Ensemble member 0 is the standard (default) CESM2–CAM6 setup. The remaining 262 ensemble members are run with parameters determined with the Latin hypercube sampling, for which the minimum and maximum values are given in Table 1. The values in Table 1 are the physical values, while the Latin hypercube sampling uses the normalized ranges to determine the parameter values. The range of values in Table 1 is chosen by “expert elicitation” among the parameterization developers for cloud microphysics, convection, unified turbulence and aerosol activation.
Some of the chosen ranges are large. However, regardless of whether the ranges of a given parameter are realistic, the specification of some a priori range is usually done from the perspective of univariate parameter variation. Because it is typically unknown how parameter perturbations will interact, when performing simultaneous perturbations, it is advantageous to consider wider ranges of parameters. These will more fully elucidate the model's ability to produce compensating errors. Compensating errors, in turn, help to indicate where independent information on individual parameters (observations, a priori theoretical or laboratory information) may be needed to independently constrain parameters and break the compensation of errors.
The parameters encompass most of the moist physical parameterizations and aerosols. This includes the unified turbulence closure (CLUBB; Golaz et al., 2002), the cloud microphysics (MG2; Gettelman and Morrison, 2015), the Modal Aerosol Model (MAM; Liu et al., 2012) and the Zhang–McFarlane deep convection scheme (ZM; Zhang and McFarlane, 1995).
A brief description of the CLUBB parameters is found in Guo et al. (2014). clubb_C2rt is the damping of scalar variances for liquid water; increasing this makes CLUBB behave closer to complete or no cloudiness (no variance) and brightens clouds. The parameters clubb_C6rt (clubb_C6thl) and clubb_C6rtb (clubb_C6htlb) are the low and high skewness of Newtonian damping of the total water flux (potential temperature flux). Decreasing these parameters tends to boost fluxes, producing a more well-mixed layer, with minor effects on cloud brightness. The low skewness especially impacts stratocumulus, while the high skewness especially impacts cumulus. Similar parameters were perturbed simultaneously, so clubb_C6rt=clubb_C6thl and clubb_C6rtb=clubb_C6htlb. clubb_C8 describes the dissipation of skewness of the vertical velocity; increasing this parameter reduces skewness, which brightens clouds. clubb_beta sets the plume widths for liquid water potential temperature and total water. An increase in clubb_beta leads to an increase in the scalar skewness. This affects liquid water and cloud fraction. clubb_c1 is the skewness of the lower side of the C1 skewness function (standard deviation of vertical velocity); increasing clubb_c1 dims clouds. clubb_c11 is the low skewness for buoyancy damping of vertical velocity. Increasing clubb_c11 brightens clouds. clubb_c14 is a constant for the dissipation of and (variances of the horizontal velocity components), and lower values brighten clouds. clubb_c_K10 is a coefficient in the momentum equation. An increase in clubb_c_K10 increases the eddy diffusivity of momentum, which, in turn, increases near-surface wind magnitude. clubb_gamma_coef controls the skewness of the vertical velocity distributions (different moments), and lowering it brightens low clouds. clubb_wpxp_L_thresh is a threshold for turbulent mixing length, below which extra damping is applied to scalar fluxes. A higher value means that the extra damping is applied to a greater range of mixing lengths.
The MG2 microphysics scheme (Gettelman and Morrison, 2015; Gettelman et al., 2015) takes bulk water and divides it into four hydrometeor categories (cloud liquid, ice, rain and snow), predicting mass and number mixing ratios for each. Several parameters are used to control the rain formation processes of autoconversion and accretion. Autoconversion is the coalescence of cloud droplets that become rain, and it is dependent on the cloud water mass mixing ratio (qd) and inversely dependent on drop number (Nd). micro_mg_autocon_lwp_exp alters the exponent on qd, and micro_mg_autocon_nd_exp alters the exponent on Nd. micro_mg_autocon_fact linearly scales autoconversion. Accretion is the process of rain drops collecting cloud water. micro_mg_accre_enhan_fact linearly scales it. micro_mg_berg_eff_factor scales the rate of vapor deposition onto ice (which also impacts supercooled liquid). micro_mg_max_nicons is the maximum allowed ice number concentration. micro_mg_dcs is the threshold diameter for cloud ice to autoconvert to snow. micro_mg_iaccr_factor similarly scales the accretion of cloud ice by snow. micro_mg_effi_factor scales the size used for the optics calculation for cloud ice. micro_mg_homog_size alters the initial size generated when a liquid homogeneously freezes to ice. Finally micro_mg_vtrmi_factor linearly scales the ice and snowfall speed.
Several parameters are related to aerosols, mostly aerosol emissions, cloud particle nucleation and scavenging. microp_aero_npccn_scale scales the activated cloud condensation nuclei (CCN) concentration, affecting drop number concentration. Subgrid-scale vertical velocities are used for both cloud droplet activation (wsub) and ice nucleation (wsubi) and are derived from the turbulent kinetic energy calculation in CLUBB. This calculation applies maximum and minimum limits to the subgrid vertical velocities. Here, we perturb the minimum values which are set with microp_aero_wsub_min and microp_aero_wsubi_min. The subgrid vertical velocities are linearly scaled with microp_aero_wsub_scale and microp_aero_wsubi_scale. Higher subgrid vertical velocities will generally activate more aerosol, leading to higher drop and crystal numbers. Dust emissions are linearly scaled with dust_emis_fact and sea-salt emissions with seasalt_emis_scale. Finally, the scavenging of aerosols in clear air below cloud by precipitation is scaled by sol_factb_interstitial and within cloud by sol_factic_interstitial.
Deep moist convection is parameterized by the ZM scheme. cldfrc_dp1 and cldfrc_dp2 define the shape of the relationship between convective mass flux and convective cloud fraction (dp1 = linear term, dp2 = log term). An increase in either of these parameters increases convective cloud fraction. The autoconversion of a convective condensate to precipitation increases by increasing zmconv_c0_lnd (over land) and zmconv_c0_ocn (over ocean), increasing the efficiency of convective precipitation. zmconv_capelmt is the convective available potential energy (CAPE) triggering threshold for deep convection, where a higher value triggers less often and allows more CAPE to build up. zmconv_dmpdz changes the entrainment rate for the initial parcel buoyancy test, and a larger value means more mixing and damped convection. zmconv_ke is the convective evaporation efficiency over ocean and zmconv_ke_lnd over land. Larger values mean more evaporation. There are two parameters for the pressure term in the convective momentum transport equation: zmconv_momcd is for downdrafts and zmconv_momcu for updrafts. Increasing them reduces the impact of resolved vertical wind shear. zmconv_num_cin is the allowed number of negative buoyancy crossings before the convective top is reached. Larger values mean deeper convection. Finally, zmconv_tiedke_add is a convective parcel temperature perturbation, where a higher value means more buoyant parcels and deeper convection.
3.4 Emulator description
We perform analysis of the raw model output across the PPE and also use several different emulation tools to analyze the ensemble. This is done to show the potential use of machine learning on the PPE. Here we focus on how well fast emulators can reproduce specific model features and then show a simple demonstration of how they can be used to tune the model. We utilize two separate emulator toolkits. The first is the Earth System Emulator (ESEm), which is an open-source tool providing a general workflow for emulating and validating a wide variety of models and outputs (Watson-Parris et al., 2021). This tool uses well-established libraries for the emulation of general circulation models with different regression techniques (neural network, Gaussian process and random forest) and provides hardware-optimized functions for efficiently sampling them. The tool also features the ability to train on 2-D fields of data. The second is a neural network emulator (hereafter referred to as the Columbia NN or Columbia emulator) developed for tuning the NASA GISS GCM (ModelE), with the final model being a combination of up to 12 different neural network models (or setups).
In our emulations, we used 210 simulations for training data (80 %) and 52 simulations for test data (20 %) with no separate samples withheld for testing. Below are longer descriptions of the emulator techniques used here.
3.4.1 Neural network
Inspired by the human brain, neural networks (NNs) form a class of flexible and expressive nonlinear functions parameterized by a large number of weights. They generally consist of multiple layers of nodes connected by edges. Each node consists of a simple (differentiable) activation function which transforms weighed input into outputs for the following nodes. The weights are optimized using gradient descent against the provided training data. The structure (or architecture) of the NN, including the number and connectivity of the layers, provides a strong inductive bias on the skill of the trained network.
The Columbia approach uses an ensemble of several NNs whose outputs are averaged. Tests showed that this methodology reduced emulator predictive noise and bias, relative to GCM output. The ensemble members are selected on the basis of minimum validation (mean square and mean absolute) error. Each NN uses a fully connected design whereby each node in each layer is connected to every node in the next layer, sometimes referred to as a multi-layer perceptron (MLP). The activation function is either a rectified linear unit (ReLU) or a leaky ReLU, depending on the NN used. The hyperparameters of each NN were chosen by manual iteration through various values of nodes-per-layer and choice of activation function (see below). The Adam optimizer was used with mean square error during training with a learning rate of 0.001. Early stopping with a patience of 100 epochs was used to prevent overfitting, using validation loss – typically, training required between 200 and 500 epochs. Most of these design choices were determined ad hoc. The Columbia emulator was originally designed to emulate the effect of GCM parameter perturbations (45 for ModelE) on the values of climatological GCM output performance scores (36 scalar diagnostics for ModelE), with skill quantified using the equivalent satellite climatologies. This emulator was not designed to output spatially resolved fields.
The ESEm NN tries to capture the spatial covariance of the full model output fields using a fully convolutional neural network (CNN). Rather than being fully connected, which would lead to a prohibitive number of parameters, CNNs convolve small kernels over the image to learn relevant features. While still requiring more parameters than assuming grid-point independence, and hence more training data, we have found that with suitable normalization such emulators can skillfully reproduce CAM6 model fields for unseen parameter combinations (see Sect. 4.3). Note also that the ESEm CNN was designed for 2-D fields and does not work as well for global averages.
3.4.2 Gaussian process emulator
A Gaussian process (GP) regression is a non-parametric approach that finds a distribution over the possible functions f(x) that are consistent with the observed data. It begins with a prior distribution and updates the prior distribution as new data points are observed, producing the posterior distribution over functions. The priors are called kernels or covariance functions. There are several different kernels that can be used, for example constant, linear and radial basis function (RBF), expressing different prior beliefs over the functional form of the model response. The kernel length scale and the smoothness parameters (sometimes referred to as hyper-parameters) can then be fit using standard optimization tools.
A key benefit of Gaussian process emulators over other approaches is that they can provide well-calibrated uncertainty quantification on their predictions. This is particularly important if the emulator is to be used for model calibration.
3.4.3 Random forest emulator
Random forest (RF) emulators generate a multitude of decision trees at the training time. The RF emulator creates several decision trees by randomly picking samples to make decisions over, reducing the risk of overfitting. A feature of this approach is that any predictions made must fall within the distribution of the training data by construction. That is to say that an RF regression model cannot extrapolate beyond the training data.
4.1 Spread across the PPE
First, we will illustrate the basic spread across the PPE for several key features of the simulated climate system. The ensembles are spread over all parameter values based upon the uniform sampling of the parameter values within the expert-chosen ranges. The magnitude of the spread in output values is dependent on the sensitivity of the parameter, the range of the parameter and the combinations together with other parameters. We first show results for a few different outputs from the three scenarios. Figure 3 shows the ensemble zonal annual mean and ±1 standard deviation (σ, shaded region) across the ensemble for aerosol optical depth (AOD; Fig. 3a), column ice water path (IWP; Fig. 3b), vertically integrated accumulation mode sulfate mass (BURDEN SO4; Fig. 3c), vertically integrated cloud condensation nuclei at 0.1 % supersaturation (CCN 0.1 %; Fig. 3d), column liquid water path (LWP; Fig. 3e), longwave cloud radiative effect (LWRE; Fig. 3f), cloud-top number concentration for liquid (ACTNL; Fig. 3g) and a histogram of the global annual mean net top-of-atmosphere (TOA) flux balance (Fig. 3h).
The zonal mean plots and histogram include all 263 members across each of the 3 run types (PD, SST4K and PI). Note that the default case (dotted line) need not be near the ensemble mean (solid line), though it is generally within ±1σ of the mean. This is not unexpected as the default parameter settings are not necessarily near the center of the range (see Table 1). Several features stand out. First, the spread of IWP (Fig. 3b) and LWP (Fig. 3e) is large – roughly a factor of 2–4. Second, the reduced BURDEN SO4, CCN and cloud-top number (Fig. 3c, d and g, respectively) in the Northern Hemisphere in the PI ensemble is clear. Also note that there is quite a spread in net TOA flux (Fig. 3h). This means a large heat gain (positive) or loss (negative) from the system. For PI and PD, most values are positive, while they are less positive for the SST4K ensemble. A stable climate is possible in these configurations with large net TOA flux because there is an unbounded source/sink of heat associated with the fixed ocean temperature, which constitutes ∼ 70 % of the surface.
4.2 Forcing and feedback
One of the unique aspects of the PPE is that in addition to the control (PD) climate, since we run the same parameter sets with perturbed climate, we can look at the variability in modeled climate responses. The differences from the PD to PI simulations are only due to aerosol emissions (greenhouse gases and SSTs remain the same). This enables us to look at the effects of anthropogenic aerosols on climate. Aerosol effects comprise both direct scattering and absorption of radiation, as well as indirect changes due to changes in cloud drop number from increased nucleation sites (Twomey, 1977) and further cloud adjustments (Albrecht, 1989; Bellouin et al., 2020). Simulations with +4 K uniformly warmer SSTs (SST4K) are commonly used to look at fast feedbacks in the atmosphere in response to surface warming (Cess et al., 1989) and have been shown to be generally similar to feedbacks with a more complete model treatment such as with a mixed-layer ocean model (e.g., Gettelman et al., 2012). To evaluate the forcing and feedback we use the weighted global mean of each ensemble member and subtract the different run types (PD−PI and SST4K−PD; Fig. 4).
Focusing on the aerosol forcing (PD−PI, blue), the difference in clear-sky TOA shortwave radiation (FSNTC; Fig. 4d) is a measure of the direct effect of aerosols and is about −0.4 W m−2. There is an increase in AOD (Fig. 4a), without much spread among the different parameter samples, and an increase in LWP (Fig. 4b), with some parameter sets producing very small increases but with a longer tail of the distribution. The ensemble average TOA flux change (Fig. 4h) is similar to the default parameter set at about −1.5 W m−2, but some sets have an aerosol forcing of lower magnitude than −1 W m−2 and some more than −2 W m−2. Given the large diversity in model state (e.g., factor of 2–4 difference in LWP and IWP, and TOA differences of up to 40 W m−2; Figs. 3b, 3e and 3g, respectively), it is remarkable that the histogram of TOA net forcing is nearly Gaussian around the default value and with a range of only −2.5 to 0 W m−2. This is close to the assessed range of aerosol forcing by Bellouin et al. (2020), although we do not explore uncertainty in absorbing aerosol (such as black carbon), which would be expected to increase the tail of uncertainty to encompass positive forcing values. Other fields are similarly distributed for PD−PI with the exception of the change in cloud drop number (Fig. 4g), which drives cloud brightening and results in cloud adjustments. Also note that changes (both large and small) in cloud radiative effects in the shortwave radiative effect (SWRE) and LWRE are nearly opposite to each other. This may be due to high-cloud changes: high clouds have large shortwave (SW) and longwave (LW) effects which are opposite, so larger LW changes would be offset by SW changes. It is still noteworthy that there is not more spread. Also, it is interesting that the cloud-top number change PD−PI has a significant spread (Fig. 4g).
For feedback (SST4K−PD) results (orange in Fig. 4), most of the distributions are slightly broader compared to the aerosol forcing. The larger magnitude of TOA difference in SST4K−PD (Fig. 4d) is likely due to the large extra heat source of emission from the warmer ocean. This is consistent with the absolute magnitude of changes in LWRE being larger in SST4K−PD than PD−PI (Fig. 4f), while the absolute magnitude for the change in SWRE is similar (Fig. 4e). There is generally a decrease in cloud fraction and an increase in outgoing clear-sky LW radiation. There is a positive change in SWRE (which has a negative magnitude; Fig. 4e) and a negative change in LWRE (which has a positive magnitude; Fig. 4f), representing a weakening of cloud radiative effect consistent with loss of clouds. In the CESM2–CAM6 PPE, every simulation loses clouds with the 4 K increase in SSTs (Fig. 4c), and almost all have the same sign of cloud changes. This is a representation of positive cloud feedbacks seen in CESM2 and other models (e.g., Zelinka et al., 2020).
4.3 Emulator results
Running climate models for large numbers of simulations can be computationally expensive. With the wealth of information from our PPE experiment, we can instead use emulators trained on the PPE data to obtain more insight into how the model behaves, how to optimize it and how scientific climate questions can be addressed. In Sect. 3.4, we described the different emulators used here. We focus on four different outputs when evaluating the emulator results: LWRE, SWRE, LWP and the residual top-of-model energy balance (RESTOM). Note that RESTOM is similar to TOA energy balance but at the top of the model as opposed to the top of the atmosphere. We emulate the response of model output to perturbations of all parameters in Table 1. Three of the emulators are from the ESEm package: CNN, GP and RF. The fourth emulator is the Columbia NN. The NN emulator was trained on 16 outputs simultaneously, while the CNN, GP and RF emulators were trained on each individual output separately. Figure 5 shows the global mean of the emulated results against the 52 PPE test ensembles, while Table 2 shows the error statistics (coefficient of determination, R2) and RMSE. Note that for this example, the Columbia NN and GP are emulated with global mean values, while the RF and CNN are emulated over the 2-D field, where the global mean is calculated after emulation. Recall that the CNN emulator is built for emulating 2-D fields and cannot be used to emulate over global means. For most of the outputs, the ESEm GP and Columbia NN emulators provide the best results. They have the highest R2 values and the lowest RMSE values (Table 1). The CNN emulator also has high R2 values; however, the RMSE values are slightly higher compared to the Columbia NN and GP emulators. The RF emulator gives the lowest score.
As stated, the ESEm tool is also able to emulate 2-D fields. Figure 6 shows an example of the 2-D results with the ESEm emulators for LWRE. Figure 6a shows the mean of the 52 test simulations, while Fig. 6b–d show the difference between the emulated results and the test simulations. Figure 6e–f show the RMSE. The total average RMSEs of LWRE along with SWRE, RESTOM and LWP are also shown in Table 3. In these cases, like when considering the global average, the GP emulator has the lowest RMSE. In this case (as opposed to GP emulation of global means in Fig. 5), the GP is emulated over the 2-D fields. However, again, we find that the GP has the best performance compared to the RF and CNN.
4.4 Sensitivity of present day (PD) climate to parameters
With the large PPE, we can evaluate which parameters have the most impact on various outputs. In the following discussions we will show results from the full PPE and the Columbia NN emulator. Along with the ESEm GP emulator, the Columbia NN emulator typically had the lowest RMSE for the global average outputs when compared with the test data as shown in Table 2. Figure 7 shows how LWRE, SWRE and RESTOM depend on values of the cloud-to-rain autoconversion exponent parameter (micro_mg_autocon_lwp_exp) and the scaling parameter for fall speed of cloud ice and snow (micro_mg_vtrmi_factor). The parameter values (x axis) are normalized (scaled by the minimum and maximum parameter values), while the output values (y axis) are standardized (scaled by the mean and standard deviation of the output values). The blue colors represent the entire PPE (263 samples), while the black and gray colors represent Columbia NN emulated results using 5000 parameter sets. By evaluating the linear regression slope of the standardized outputs, the parameters with the largest slopes (in absolute terms) are determined to have the largest impact on the outputs. Since the outputs are standardized and parameters are normalized, the slopes for different outputs and parameters are directly comparable. Slopes can be calculated for all outputs and parameters. For the example in Fig. 7, it is clear that the cloud ice particle fall speed parameter has a large impact on LWRE, while the autoconversion parameter value is important for RESTOM and SWRE. In the cases shown here, the regression slope from the PPE ensemble and the regression slope from the emulated results are almost identical. This indicates that the emulator can reproduce the spread of the PPE well. We note that for most parameter and output sets we produced, the emulator reproduced well the PPE regression slope (not shown). Note that with this simple evaluation we can obtain the regression slope directly from the PPE. Thus, a full emulation to obtain the regression slope is not necessary, and the emulation results are included here primarily to illustrate the performance of the emulator. We also acknowledge that the assumption that the outputs change linearly with the parameters is not necessarily true in all instances; however, this assumption is reasonable for an initial evaluation. Furthermore, when looking at the coefficient of determination, it is evident that higher values are correlated with steeper slopes (not shown).
Figure 8 shows a grid plot of the linear regression slopes (lines in Fig. 7) for 16 outputs (vertical axis) against the 43 parameter values (horizontal axis). Blue values mean that the output decreases with increasing parameter value, and red means that the output increases with parameter value. The darker the colors are, the steeper the slope is and the more the output is dependent on the parameter value. Since weather and climate systems can be different in different regions, we show global results as well as results from the Arctic, midlatitudes, tropics and the Southern Ocean. The vertical black lines divide the parameters into their respective physics package: deep convection, aerosol, microphysics and turbulence. The parameters are listed in the same order as in Table 1. The outputs (vertical axis going down) are listed in the order of radiation, cloud properties and aerosol properties.
Some parameters stand out in almost all regions for many of the outputs, especially the microphysical parameters. The accretion enhancement factor (micro_mg_accre_enhan_fact), the autoconversion scaling factor (micro_mg_autocon_fact) and the autoconversion exponent (micro_mg_autocon_lwp_exp) all directly affect rain formation and the amount of liquid water in the atmosphere. The autoconversion size threshold of cloud ice to snow (micro_mg_dcs) and the ice sedimentation factor (micro_mg_vtrmi_factor) strongly influence the ice water path. However, in the tropics, where the deep convection scheme has a dominant influence, only the micro_mg_vtrmi_factor remains as an important microphysical parameter for most of the outputs. Nonetheless, the parameter micro_mg_dcs is still important for the LWRE in the tropics. This is not surprising since it has a large impact on cirrus clouds.
The deep convection parameter that impacts radiation outputs the most is the convective parcel temperature perturbation (zmconv_tiedke_add), and this is especially true in the tropics. The triggering threshold for convection (zmconv_capelmt) affects the ice water path and the sulfate burden. These may be related through sulfate effects on the homogeneous nucleation of ice. Aerosol parameters have less impact on radiation outputs, but several of them are important for cloud properties and precipitation. The turbulence parameters have a relatively lesser impact on the outputs presented here compared to the microphysics, aerosol and convection parameters. This might be because the selected range for some of the key CLUBB parameters (like clubb_gamma_coef) is narrower compared to those used by others. For example, other PPE approaches with different versions of CAM have found the shallow cloud turbulence to be important (Guo et al., 2015) with a broader range of some CLUBB parameters.
4.5 Tuning example
One of the goals of PPE studies is to assist with constraining (“tuning”) parameters in models. Though this is not the main goal of this paper, we experimented with tuning the CESM2–CAM6 model against the Clouds and the Earth's Radiant Energy System (CERES) and Multisensor Advanced Climatology of Liquid Water Path (MAC-LWP; Elsaesser et al., 2017) products using the Columbia NN emulator. To obtain enough samples we used 20 000 000 parameter samples with the emulator, creating the parameter samples using the Latin hypercube sampling technique with all parameters normally distributed. For tuning, we focus on LWRE, SWRE, RESTOM and LWP, and Fig. 9 shows distributions of the emulated outputs. The targets are the observed global means (RESTOM: 0 W m−2; SWRE: −45.2 W m−2; LWRE: 25.7 W m−2; LWP: 0.065 kg m−2), indicated by the solid red lines in Fig. 9. CAM6 output only includes the “cloud” portion of LWP, and since observations cannot easily distinguish cloud from rainwater LWP, we followed the recommendation of Elsaesser et al. (2017) and used only grid boxes for which the ratio of observed cloud to total LWP exceeded 0.8 in the global mean LWP calculation for both MAC-LWP and CAM6. We look for all emulated outputs that are within the CERES mean ±2 W m−2 for LWRE, ±3 W m−2 for SWRE and RESTOM, and ±0.01 kg m−2 for LWP. The ranges are chosen to allow for enough samples to fall within the ranges in order to produce meaningful probability density functions (PDFs) in Fig. 10, while these values could be set to correspond to observational or emulator uncertainties. All parameter sets that are within the range for all four outputs are accepted; this effectively defines a bounded uniform likelihood over the 4-D observational space. Figure 10 shows the histograms of the parameter values that result in outputs within the selected ranges. The solid red lines in Fig. 10 indicate the default parameter value in CESM2–CAM6.
For parameter histograms that are non-uniform and strongly peaked, relatively more samples near the peak are within the acceptable tuning range. For example, the clubb_C8 parameter peaks at relatively large values, while the clubb_c14 parameter peaks at small values. But again, this result might be due to the relatively narrower chosen range for CLUBB parameters compared to the other types of parameters. Also interesting is the fact that a relatively low dust emission factor but high sea salt emission factor most often produce outputs consistent with the observations within the acceptable range. We emphasize that parameters with a peaked histogram in Fig. 10 are not necessarily the parameters for which the outputs are most sensitive as determined by the regression slope magnitudes (outputs regressed on parameter values) in Fig. 8. For instance, the dust and sea salt emission factors have strongly peaked histograms giving outputs in the acceptable tuning range but relatively small regression slopes (Fig. 8). This apparent discrepancy can be explained by the linear nature of regression versus the nonlinear emulator. There is evidence that there is a complicated relationship between different parameters.
Ice fall speed (micro_mg_vtrmi_factor) and the number of levels of convective inhibition in the deep convection scheme (zmconv_num_cin) are the only parameters with a strong peak in the middle of their range. As previously stated, the ranges were chosen by expert elicitation with the default values within the minimum and maximum range. However, the values giving realistic outputs here are most often near the edge of the physically plausible parameter ranges as determined by expert guidance. Some of the parameters, such as clubb_C8 and clubb_C11, have default settings near the upper end of the ranges which are close to the histogram peaks. Other parameters, such as clubb_c14 and dust_emis_fact, have default settings near the middle of the range but strongly peaked histograms at the edge. Thus, these parameters most often have values at the edge of their range, different from the default values, to produce outputs consistent with observations. One possible explanation for this behavior could be that there are structural errors in the model, and thus we must push some parameter values to the edge of their plausible range to obtain results consistent with observations. On the other hand, several parameters are fairly uniform over the entire parameter range, such as for example clubb_c1 and zmconv_tiedke_add. These results indicate that any value of these parameters within their range can produce outputs close to observations. It is possible that by considering more observational targets, these parameters could be further constrained and unreasonable parameter combinations could be eliminated.
Here we have presented a CESM2–CAM6 perturbed parameter ensemble PPE. We perturbed 45 parameters in the microphysics, turbulence, deep convection and aerosol physics packages and generated an ensemble with 263 members. Simulations were generated for current climate, preindustrial aerosol loading and future climate with 4 K added to the sea surface temperature. The main objective of this paper is to provide a description of the CESM2–CAM6 PPE dataset and present some initial results. The main results can be summarized as follows:
-
The PPE has many different usages, for example, understanding uncertainties in model parameterizations, climate sensitivities to parameter values and optimal parameter tuning. The CESM2–CAM6 PPE data are publicly available for the community to use. The CESM2–CAM6 PPE is extensible, and new PPE datasets can be created in a straightforward way using other parameter combinations or different model setups.
-
Of the outputs evaluated here, there is a large spread in IWP, LWP and TOA among the individual ensembles (Fig. 3). Large TOA fluxes in many ensemble members are possible only because with fixed SSTs there is an unbounded heat source/sink at the ocean surface that stabilizes the climate. Large ranges in LWP and IWP indicate that some parameters can significantly increase or decrease cloud cover, although there is more constraint on the radiative fluxes since the radiative forcing is nonlinear with respect to cloud mass.
-
Both aerosol forcing (PD−PI) and cloud feedback (SST4K−PD) show a spread in the output values considered here (Fig. 4). However, the aerosol forcing range is relatively narrow compared to the cloud feedback (except for the cloud-top number concentration). There is more spread in the total cloud cover and LWP in the cloud feedback case. This drives the TOA differences as the cloud environment varies more with SST4K.
-
We tested various emulators that were applied to the PPE ensemble. The Columbia NN, ESEm GP emulator and ESEm CNN all produce reasonable results for selected outputs, while the ESEm RF emulator had the lowest scores when considering global means. Both the CNN and RF outputs were emulated on 2-D fields, while the error statistics were calculated on the global mean values. When calculating the error statistics on the 2-D fields, the RF performed at times better than the CNN emulator, while the GP emulator still had the best score overall.
-
With the large number of parameters, we evaluated the sensitivity of global outputs when changing the parameter values. There were a select number of parameters that have strong sensitivity, especially several microphysics parameters. The pattern changes slightly when considering specific zonal regions, such as the Arctic, midlatitudes, tropics and the Southern Ocean. For example, the microphysics parameters create higher sensitivity in the Arctic and midlatitudes than in the tropics and Southern Ocean, while some deep convection parameters have more impact in the tropics and Southern Ocean.
-
We provided a simple tuning experiment using the Columbia NN emulator. We identified the parameter combinations that gave results within a small range of observed global values and evaluated distributions of parameter values from these combinations. A few parameter distributions peak within the range of physically plausible parameter values (as determined by expert guidance), while several parameters peak at the edge of the parameter ranges. Only four observational targets were used to constrain parameter values. By including more observations, parameters may be better constrained. Furthermore, more elaborate techniques for sampling constrained parameter values, such as Markov chain Monte Carlo and other Bayesian approaches, could improve the efficiency and accuracy of tuning and allow for a more comprehensive account of observational uncertainties.
The PPE dataset and the CESM2–CAM6 code version cam6_3_026 are available at the Climate Data Gateway at NCAR (https://doi.org/10.26024/bzne-yf09, Eidhammer et al., 2022). The current version of CESM is available from https://github.com/ESCOMP/CESM (last access: 30 October 2024) under the license found here: https://www.cesm.ucar.edu/models/cesm2/copyright (last access: 30 October 2024). The specific CERES data used in this paper are available on Zenodo (https://doi.org/10.5281/zenodo.10426438, Eidhammer and Gettelman, 2024). The MAC-LWP data are available at the Goddard Earth Sciences Data and Information Services Center (https://doi.org/10.5067/MEASURES/MACLWPM, Elsaesser et al., 2016).
TE conducted the simulations to create the PPE, conducted analysis of the data and wrote the manuscript. AG initiated the idea of the CESM PPE, oversaw the production of the PPE, conducted analysis of the data and assisted in writing the manuscript. KTC generated the scripts to run the PPE simulations and assisted in editing the manuscript. DWP provided the ESEm emulators and provided input on the manuscript. GE and MvLW provided the Columbia emulator. HM initiated the idea of the CESM PPE, provided input on the work and edited the manuscript. DM and CS provided computing resources and input on the manuscript.
The contact author has declared that none of the authors has any competing interests.
Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors.
This material is based upon work supported by the NSF National Center for Atmospheric Research, which is a major facility sponsored by the U.S. National Science Foundation under cooperative agreement no. 1852977. We would like to acknowledge high-performance computing support from Cheyenne (https://doi.org/10.5065/D6RX99HX) provided by NSF NCAR's Computational and Information Systems Laboratory (2019) including through the Wyoming–NSF NCAR alliance large allocation WYOM0124.
This research has been supported by the National Aeronautics and Space Administration (grant nos. 80NSSC17K0073 and 80NSSC21K1499) and the NSF STC Learning the Earth with Artificial Intelligence and Physics (LEAP; NSF award number 2019625).
This paper was edited by Nina Crnivec and reviewed by two anonymous referees.
Albrecht, B. A.: Aerosols, Cloud Microphysics, and Fractional Cloudiness, Science, 245, 1227–1230, https://doi.org/10.1126/science.245.4923.1227, 1989. a
Bellouin, N., Quaas, J., Gryspeerdt, E., Kinne, S., Stier, P., Watson-Parris, D., Boucher, O., Carslaw, K. S., Christensen, M., Daniau, A.-L., Dufresne, J.-L., Feingold, G., Fiedler, S., Forster, P., Gettelman, A., Haywood, J. M., Lohmann, U., Malavelle, F., Mauritsen, T., McCoy, D. T., Myhre, G., Mülmenstädt, J., Neubauer, D., Possner, A., Rugenstein, M., Sato, Y., Schulz, M., Schwartz, S. E., Sourdeval, O., Storelvmo, T., Toll, V., Winker, D., and Stevens, B.: Bounding Global Aerosol Radiative Forcing of Climate Change, Rev. Geophys., 58, e2019RG000660, https://doi.org/10.1029/2019RG000660, 2020. a, b
Bogenschutz, P. A., Gettelman, A., Morrison, H., Larson, V. E., Craig, C., and Schanen, D. P.: Higher-Order Turbulence Closure and Its Impact on Climate Simulation in the Community Atmosphere Model, J. Climate, 26, 9655–9676, https://doi.org/10.1175/JCLI-D-13-00075.1, 2013. a
Bogenschutz, P. A., Gettelman, A., Hannay, C., Larson, V. E., Neale, R. B., Craig, C., and Chen, C.-C.: The path to CAM6: coupled simulations with CAM5.4 and CAM5.5, Geosci. Model Dev., 11, 235–255, https://doi.org/10.5194/gmd-11-235-2018, 2018. a
Cess, R. D., Potter, G. L., Blanchet, J. P., Boer, G. J., Ghan, S. J., Kiehl, J. T., Le Treut, H., Li, Z.-X., Liang, X.-Z., Mitchell, J. F. B.. Morcrette, J.-J., Randall, D. A., Riches, M. R., Roeckner, E., Schlese, U., Slingo, A., Taylor, K. E., Washington, W. M., Wetherald, R. T., and Yagai, I.: Interpretation of Cloud-Climate Feedback as Produced by 14 Atmospheric General Circulation Models, Science, 245, 513–516, https://doi.org/10.1126/science.245.4917.513, 1989. a
Computational and Information Systems Laboratory: Cheyenne: HPE/SGI ICE XA System (NCAR Community Computing), Boulder, CO, National Center for Atmospheric Research, https://doi.org/10.5065/D6RX99HX, 2019. a
Danabasoglu, G., Lamarque, J.-F., Bacmeister, J., Bailey, D. A., DuVivier, A. K., Edwards, J., Emmons, L. K., Fasullo, J., Garcia, R., Gettelman, A., Hannay, C., Holland, M. M., Large, W. G., Lauritzen, P. H., Lawrence, D. M., Lenaerts, J. T. M., Lindsay, K., Lipscomb, W. H., Mills, M. J., Neale, R., Oleson, K. W., Otto-Bliesner, B., Phillips, A. S., Sacks, W., Tilmes, S., van Kampenhout, L., Vertenstein, M., Bertini, A., Dennis, J., Deser, C., Fischer, C., Fox-Kemper, B., Kay, J. E., Kinnison, D., Kushner, P. J., Larson, V. E., Long, M. C., Mickelson, S., Moore, J. K., Nienhouse, E., Polvani, L., Rasch, P. J., and Strand, W. G.: The Community Earth System Model Version 2 (CESM2), J. Adv. Model. Earth Sy., 12, e2019MS001916, https://doi.org/10.1029/2019MS001916, 2020. a, b
Duffy, M. L., Medeiros, B., Gettelman, A., and Eidhammer, T.: Perturbing Parameters to Understand Cloud Contributions to Climate Change, J. Climate, 37, 213–227, https://doi.org/10.1175/JCLI-D-23-0250.1, 2024. a
Eidhammer, T. and Gettelman, A.: CERES_EBAF_Ed4.1_2001-2020 subset, Zenodo [data set], https://doi.org/10.5281/zenodo.10426438, 2024. a
Eidhammer, T., Gettelman, A., and Thayer-Calder, K.: CESM2.2-CAM6 Perturbed Parameter Ensemble (PPE), NCAR [data set], https://doi.org/10.26024/bzne-yf09, 2022. a, b
Elsaesser, G., O'Dell, C., Lebsock, M., and Teixeira, J.: Multisensor Advanced Climatology Mean Liquid Water Path L3 Monthly 1 degree x 1 degree V1, Greenbelt, MD, USA, Goddard Earth Sciences Data and Information Services Center (GES DISC) [data set], https://doi.org/10.5067/MEASURES/MACLWPM, 2016. a
Elsaesser, G. S., O’Dell, C. W., Lebsock, M. D., Bennartz, R., Greenwald, T. J., and Wentz, F. J.: The Multisensor Advanced Climatology of Liquid Water Path (MAC-LWP), J. Climate, 30, 10193–10210, https://doi.org/10.1175/JCLI-D-16-0902.1, 2017. a, b
Gettelman, A. and Morrison, H.: Advanced Two-Moment Bulk Microphysics for Global Models. Part I: Off-Line Tests and Comparison with Other Schemes, J. Climate, 28, 1268–1287, https://doi.org/10.1175/JCLI-D-14-00102.1, 2015. a, b, c
Gettelman, A., Kay, J. E., and Shell, K. M.: The Evolution of Climate Feedbacks in the Community Atmosphere Model, J. Climate, 25, 1453–1469, https://doi.org/10.1175/JCLI-D-11-00197.1, 2012. a
Gettelman, A., Morrison, H., Santos, S., Bogenschutz, P., and Caldwell, P. M.: Advanced Two-Moment Bulk Microphysics for Global Models. Part II: Global Model Solutions and Aerosol–Cloud Interactions, J. Climate, 28, 1288–1307, https://doi.org/10.1175/JCLI-D-14-00103.1, 2015. a, b
Gettelman, A., Hannay, C., Bacmeister, J. T., Neale, R. B., Pendergrass, A. G., Danabasoglu, G., Lamarque, J.-F., Fasullo, J. T., Bailey, D. A., Lawrence, D. M., and Mills, M. J.: High Climate Sensitivity in the Community Earth System Model Version 2 (CESM2), Geophys. Res. Lett., 46, 8329–8337, https://doi.org/10.1029/2019GL083978, 2019. a
Gettelman, A., Eidhammer, T., Duffy, M. L., McCoy, D. T., Song, C., and Watson-Parris, D.: The Interaction Between Climate Forcing and Feedbacks, J. Geophys. Res.-Atmos., 129, e2024JD040857, https://doi.org/10.1029/2024JD040857, 2024. a
Golaz, J.-C., Larson, V. E., and Cotton, W. R.: A PDF-Based Model for Boundary Layer Clouds. Part I: Method and Model Description, J. Atmos. Sci., 59, 3540–3551, https://doi.org/10.1175/1520-0469(2002)059<3540:APBMFB>2.0.CO;2, 2002. a, b
Guo, Z., Wang, M., Qian, Y., Larson, V. E., Ghan, S., Ovchinnikov, M., Bogenschutz, P. A., Zhao, C., Lin, G., and Zhou, T.: A sensitivity analysis of cloud properties to CLUBB parameters in the single-column Community Atmosphere Model (SCAM5), J. Adv. Model. Earth Sy., 6, 829–858, https://doi.org/10.1002/2014MS000315, 2014. a
Guo, Z., Wang, M., Qian, Y., Larson, V. E., Ghan, S., Ovchinnikov, M., A. Bogenschutz, P., Gettelman, A., and Zhou, T.: Parametric behaviors of CLUBB in simulations of low clouds in the Community Atmosphere Model (CAM), J. Adv. Model. Earth Sy., 7, 1005–1025, https://doi.org/10.1002/2014MS000405, 2015. a
Hoose, C., Kristjánsson, J. E., Chen, J.-P., and Hazra, A.: A Classical-Theory-Based Parameterization of Heterogeneous Ice Nucleation by Mineral Dust, Soot, and Biological Particles in a Global Climate Model, J. Atmos. Sci., 67, 2483–2503, https://doi.org/10.1175/2010JAS3425.1, 2010. a
Hourdin, F., Mauritsen, T., Gettelman, A., Golaz, J.-C., Balaji, V., Duan, Q., Folini, D., Ji, D., Klocke, D., Qian, Y., Rauser, F., Rio, C., Tomassini, L., Watanabe, M., and Williamson, D.: The Art and Science of Climate Model Tuning, B. Am. Meteorol. Soc., 98, 589–602, https://doi.org/10.1175/BAMS-D-15-00135.1, 2016. a
Jackson, C., Sen, M. K., Huerta, G., Deng, Y., and Bowman, K. P.: Error Reduction and Convergence in Climate Prediction, J. Climate, 21, 6698–6709, 2008. a
Lee, L. A., Carslaw, K. S., Pringle, K. J., Mann, G. W., and Spracklen, D. V.: Emulation of a complex global aerosol model to quantify sensitivity to uncertain parameters, Atmos. Chem. Phys., 11, 12253–12273, https://doi.org/10.5194/acp-11-12253-2011, 2011. a
Lee, L. A., Reddington, C. L., and Carslaw, K. S.: On the Relationship between Aerosol Model Uncertainty and Radiative Forcing Uncertainty, P. Natl. Acad. Sci. USA, 113, 5820–5827, https://doi.org/10.1073/pnas.1507050113, 2016. a
Liu, X., Easter, R. C., Ghan, S. J., Zaveri, R., Rasch, P., Shi, X., Lamarque, J.-F., Gettelman, A., Morrison, H., Vitt, F., Conley, A., Park, S., Neale, R., Hannay, C., Ekman, A. M. L., Hess, P., Mahowald, N., Collins, W., Iacono, M. J., Bretherton, C. S., Flanner, M. G., and Mitchell, D.: Toward a minimal representation of aerosols in climate models: description and evaluation in the Community Atmosphere Model CAM5, Geosci. Model Dev., 5, 709–739, https://doi.org/10.5194/gmd-5-709-2012, 2012. a, b
McKay, M. D., Beckman, R. J., and Conover, W. J.: A comparison of three methods for selecting values of input variables in the analysis of output from a computer code, Technometrics, 42, 55–61, 2000. a
Mills, M. J., Schmidt, A., Easter, R., Solomon, S., Kinnison, D. E., Ghan, S. J., Neely, R. R., Marsh, D. R., Conley, A., Bardeen, C. G., and Gettelman, A.: Global Volcanic Aerosol Properties Derived from Emissions, 1990–2014, Using CESM1(WACCM), J. Geophys. Res.-Atmos., 121, 2015JD024290, https://doi.org/10.1002/2015JD024290, 2016. a
Morrison, H. and Gettelman, A.: A new two-moment bulk stratiform cloud microphysics scheme in the Community Atmosphere Model, version 3 (CAM3). Part I: Description and numerical tests, J. Climate, 21, 3642–3659, 2008. a
Morrison, H., van Lier-Walqui, M., Fridlind, A. M., Grabowski, W. W., Harrington, J. Y., Hoose, C., Korolev, A., Kumjian, M. R., Milbrandt, J. A., Pawlowska, H., Posselt, D. J., Prat, O. P., Reimel, K. J., Shima, S.-I., van Diedenhoven, B., and Xue, L.: Confronting the Challenge of Modeling Cloud and Precipitation Microphysics, J. Adv. Model. Earth Sy., 12, e2019MS001689, https://doi.org/10.1029/2019MS001689, 2020. a
Peatier, S., Sanderson, B. M., Terray, L., and Roehrig, R.: Investigating Parametric Dependence of Climate Feedbacks in the Atmospheric Component of CNRM-CM6-1, Geophys. Res. Lett., 49, e2021GL095084, https://doi.org/10.1029/2021GL095084, e2021GL095084 2021GL095084, 2022. a
Posselt, D. J. and Vukicevic, T.: Robust Characterization of Model Physics Uncertainty for Simulations of Deep Moist Convection, Mon. Weather Rev., 138, 1513–1535, 2010. a
Qian, Y., Yan, H., Hou, Z., Johannesson, G., Klein, S., Lucas, D., Neale, R., Rasch, P., Swiler, L., Tannahill, J., Wang, H., Wang, M., and Zhao, C.: Parametric sensitivity analysis of precipitation at global and local scales in the Community Atmosphere Model CAM5, J. Adv. Model. Earth Sy., 7, 382–411, https://doi.org/10.1002/2014MS000354, 2015. a, b
Qian, Y., Wan, H., Yang, B., Golaz, J.-C., Harrop, B., Hou, Z., Larson, V. E., Leung, L. R., Lin, G., Lin, W., Ma, P.-L., Ma, H.-Y., Rasch, P., Singh, B., Wang, H., Xie, S., and Zhang, K.: Parametric Sensitivity and Uncertainty Quantification in the Version 1 of E3SM Atmosphere Model Based on Short Perturbed Parameter Ensemble Simulations, J. Geophys. Res.-Atmos., 123, 13046–13073, https://doi.org/10.1029/2018JD028927, 2018. a
Regayre, L. A., Pringle, K. J., Booth, B. B. B., Lee, L. A., Mann, G. W., Browse, J., Woodhouse, M. T., Rap, A., Reddington, C. L., and Carslaw, K. S.: Uncertainty in the magnitude of aerosol-cloud radiative forcing over recent decades, Geophys. Res. Lett., 41, 9040–9049, https://doi.org/10.1002/2014GL062029, 2014. a
Regayre, L. A., Johnson, J. S., Yoshioka, M., Pringle, K. J., Sexton, D. M. H., Booth, B. B. B., Lee, L. A., Bellouin, N., and Carslaw, K. S.: Aerosol and physical atmosphere model parameters are both important sources of uncertainty in aerosol ERF, Atmos. Chem. Phys., 18, 9975–10006, https://doi.org/10.5194/acp-18-9975-2018, 2018. a, b
Regayre, L. A., Deaconu, L., Grosvenor, D. P., Sexton, D. M. H., Symonds, C., Langton, T., Watson-Paris, D., Mulcahy, J. P., Pringle, K. J., Richardson, M., Johnson, J. S., Rostron, J. W., Gordon, H., Lister, G., Stier, P., and Carslaw, K. S.: Identifying climate model structural inconsistencies allows for tight constraint of aerosol radiative forcing, Atmos. Chem. Phys., 23, 8749–8768, https://doi.org/10.5194/acp-23-8749-2023, 2023. a
Schmidt, G. A., Bader, D., Donner, L. J., Elsaesser, G. S., Golaz, J.-C., Hannay, C., Molod, A., Neale, R. B., and Saha, S.: Practice and philosophy of climate model tuning across six US modeling centers, Geosci. Model Dev., 10, 3207–3223, https://doi.org/10.5194/gmd-10-3207-2017, 2017. a
Song, C., McCoy, D. T., Eidhammer, T., Gettelman, A., McCoy, I. L., Watson-Parris, D., Wall, C. J., Elsaesser, G., and Wood, R.: Buffering of Aerosol-Cloud Adjustments by Coupling Between Radiative Susceptibility and Precipitation Efficiency, Geophys. Res. Lett., 51, e2024GL108663, https://doi.org/10.1029/2024GL108663, 2024. a
Trenberth, K. E. and Fasullo, J. T.: Simulation of present-day and twenty-first-century energy budgets of the southern oceans, J. Climate, 23, 440–454, 2010. a
Twomey, S.: The influence of pollution on the shortwave albedo of clouds, J. Atmos. Sci., 34, 1149–1152, 1977. a
van Lier-Walqui, M., Vukicevic, T., and Posselt, D. J.: Quantification of Cloud Microphysical Parameterization Uncertainty using Radar Reflectivity, Mon. Weather Rev., 140, 3442–3466, 2012. a
Wagman, B. M. and Jackson, C. S.: A Test of Emergent Constraints on Cloud Feedback and Climate Sensitivity Using a Calibrated Single-Model Ensemble, J. Climate, 31, 7515–7532, https://doi.org/10.1175/JCLI-D-17-0682.1, 2018. a
Watson-Parris, D., Bellouin, N., Deaconu, L., Schutgens, N. A., Yoshioka, M., Regayre, L. A., Pringle, K. J., Johnson, J. S., Smith, C., Carslaw, K., and Stier, P.: Constraining uncertainty in aerosol direct forcing, Geophys. Res. Lett., 47, e2020GL087141, https://doi.org/10.1029/2020GL087141, 2020. a
Watson-Parris, D., Williams, A., Deaconu, L., and Stier, P.: Model calibration using ESEm v1.1.0 – an open, scalable Earth system emulator, Geosci. Model Dev., 14, 7659–7672, https://doi.org/10.5194/gmd-14-7659-2021, 2021. a
Zelinka, M. D., Myers, T. A., McCoy, D. T., Po-Chedley, S., Caldwell, P. M., Ceppi, P., Klein, S. A., and Taylor, K. E.: Causes of higher climate sensitivity in CMIP6 models, Geophys. Res. Lett., 47, e2019GL085782, https://doi.org/10.1029/2019GL085782, 2020. a
Zhang, G. J. and McFarlane, N. A.: Sensitivity of climate simulations to the parameterization of cumulus convection in the Canadian Climate Centre general circulation model, Atmos. Ocean, 33, 407–446, 1995. a, b