A new exponentially decaying error correlation model for assimilating OCO-2 column-average CO2 data using a length scale computed from airborne lidar measurements
- 1Cooperative Institute for Research in the Atmosphere, Colorado State University, Fort Collins, CO, USA
- 2Department of Meteorology and Atmospheric Science, Pennsylvania State University, University Park, PA, USA
- 3Earth and Environmental Systems Institute, Pennsylvania State University, University Park, PA, USA
- 4NASA Langley Research Center, Hampton, VA, USA
- 5Spectral Sensor Solutions LLC, Fort Wayne, IN, USA
Correspondence: David F. Baker (email@example.com)
To check the accuracy of column-average dry air CO2 mole fractions () retrieved from Orbiting Carbon Observatory (OCO-2) data, a similar quantity has been measured from the Multi-functional Fiber Laser Lidar (MFLL) aboard aircraft flying underneath OCO-2 as part of the Atmospheric Carbon and Transport (ACT) – America flight campaigns. Here we do a lagged correlation analysis of these MFLL–OCO-2 column CO2 differences and find that their correlation spectrum falls off rapidly at along-track separation distances under 10 km, with a correlation length scale of about 10 km, and less rapidly at longer separation distances, with a correlation length scale of about 20 km.
The OCO-2 satellite takes many CO2 measurements with small (∼3 km2) fields of view (FOVs) in a thin (<10 km wide) swath running parallel to its orbit: up to 24 separate FOVs may be obtained per second (across a ∼6.75 km distance on the ground), though clouds, aerosols, and other factors cause considerable data dropout. Errors in the CO2 retrieval method have long been thought to be correlated at these fine scales, and methods to account for these when assimilating these data into top-down atmospheric CO2 flux inversions have been developed. A common approach has been to average the data at coarser scales (e.g., in 10 s long bins) along-track, then assign an uncertainty to the averaged value that accounts for the error correlations. Here we outline the methods used up to now for computing these 10 s averages and their uncertainties, including the constant-correlation-with-distance error model that was used to summarize the OCO-2 version 9 retrievals as part of the OCO-2 flux inversion model intercomparison project. We then derive a new one-dimensional error model using correlations that decay exponentially with separation distance, apply this model to the OCO-2 data using the correlation length scales derived from the MFLL–OCO-2 differences, and compare the results (for both the average and its uncertainty) to those given by the current constant correlation error model. To implement this new model, the data are averaged first across 2 s spans to collapse the cross-track distribution of the real data onto the 1-D path assumed by the new model. Considering correlated errors can cause the average value to fall outside the range of the values averaged; two strategies for preventing this are presented. The correlation lengths over the ocean, which the land-based MFLL data do not clarify, are assumed to be twice those over the land.
The new correlation model gives 10 s averages that are only a few tenths of 1 ppm different from the constant correlation model. Over land, the uncertainties in the mean are also similar, suggesting that the +0.3 constant correlation coefficient currently used in the model there is accurate. Over the oceans, the twice-the-land correlation lengths that we assume here result in a significantly lower uncertainty on the mean than the +0.6 constant correlation currently gives – measurements similar to the MFLL ones are needed over the oceans to do better. Finally, we show how our 1-D exponential error correlation model may be used to account for correlations in inversion methods that choose to assimilate each retrieval individually and also to account for correlations between separate 10 s averages when these are assimilated instead.
Column-averaged CO2 mixing ratio measurements taken from satellites provide coverage across the globe that is far more extensive than that from in situ measurements. These satellite measurements are often used in global atmospheric flux inversions to provide a “top-down” constraint on surface sources and sinks of CO2. The atmospheric transport models underlying these global inversions are generally run at a coarse resolution using grid boxes of hundreds of kilometers on a side. The resolution is limited for computational reasons (the models must be run many dozens of times across the measurements to obtain the inverse estimate) and because the spatial coverage of the satellite measurements is currently not dense enough to resolve spatial scales much finer than this when solving at typical timescales (the gap in longitude between subsequent passes of a typical low-Earth-orbiting (LEO) satellite taking a single thin swath of data along its orbit path is ∼25∘, resulting in gaps of between 3 and 4∘ across a week, gaps which are generally never filled in further due to the repeat cycle of the satellite's orbit). The typical field of view (FOV) of individual retrievals is often much smaller than this grid box scale, however: FOVs for retrievals from the Orbiting Carbon Observatory (OCO-2) satellite (Crisp et al., 2004, 2008), for example, are typically ∼2.25 km along-track by at most 1.25 km across-track (Eldering et al., 2017). The individual OCO-2 retrievals are generally averaged together along-track across some distance closer to the model grid box size before being assimilated in the inversion: this is because the modeled measurements to which the true measurements will be compared in the inversion are available only at the grid box resolution, so it makes little sense to assimilate each measurement individually when assimilating a coarse-resolution average that summarizes those values will do just as well.
Whether the individual OCO-2 retrievals or coarser-resolution averages of them are assimilated into the inverse model, correlations between the errors in the individual CO2 measurements must be considered. CO2 mixing ratios in the upper part of the atmospheric column (at all levels but the immediate surface layer) feel the influence of multiple flux locations at the surface due to atmospheric mixing, which widens and homogenizes the zone of influence as time goes on. is a measure of CO2 across the full column and is dominated by such effects: any error in will be translated into highly correlated errors in neighboring surface fluxes when used as a measurement in an inversion model; similarly, any error in surface CO2 flux in a forward model will result in highly correlated errors in neighboring measurements influenced by these fluxes. Also, systematic errors in the individual CO2 retrievals are correlated at finer scales because incorrect assumptions are made about the scatterers, water vapor, temperature, and surface properties used in the retrieval scheme, and these variables themselves have errors that are correlated at these scales. The OCO-2 satellite, for example, makes 24 separate observations per second across a distance of ∼6.75 km along the ground track: these data provide mostly redundant column CO2 information across that time. When deciding how to weight the satellite data in the inversions relative to the a priori information, some assumptions about these measurement error correlations must therefore be made: if these errors were assumed to all be independent, the total amount of measurement information going into the inversions would be much too high, resulting in improper weighting versus the a priori or dynamically propagated information in the problem. Our goal here is to present a new model of the errors in the OCO-2 CO2 measurements that assumes correlations that die off exponentially with distance, as opposed to the constant correlation models used previously. This new model will allow the CO2 measurements to be weighted more accurately in inversions, yielding more accurate CO2 flux estimates.
Until recently, there have not been any good ground-truth data available to evaluate satellite CO2 measurement errors at finer scales. At coarser scales, data from the Total Carbon Column Observing Network (TCCON) have been used to assess the magnitude and seasonal variability of OCO-2 errors (Wunch et al., 2017), as well as what portion of these might be considered random as opposed to systematic (Kulawik et al., 2019). Worden et al. (2017) have done a similar random versus systematic partitioning of OCO-2 errors by looking at the variability of retrieved across small areas inside which the real CO2 is thought not to vary much. Comparisons to realistic CO2 fields given by atmospheric transport models (e.g., using plausible prior flux estimates and forced to agree with the available in situ CO2 measurements) have also been used to assess systematic errors in the satellite retrievals (O'Dell et al., 2012, 2018) at seasonal timescales and regional spatial scales. The TCCON sites provide column-averaged CO2 measurements that can be compared directly to the OCO-2 column measurements, but because they are available only in a few fixed locations, they cannot assess how these errors vary along-track. There have been many aircraft underflights of OCO-2, but these generally have taken only in situ measurements of CO2 representative of a particular elevation rather than a column average. Some of these flights provide data across most of the atmospheric column (i.e., vertical profiles) but only generally at widely spaced locations. None of these data are really well-suited for assessing along-track errors in the column average.
Over the past several years, however, column-average measurements of CO2 from aircraft-based lidars have become available. Several of these lidars were instrument test beds developed as part of NASA's Active Sensing of CO2 Emissions over Nights, Days, and Seasons (ASCENDS) satellite project (Jucks et al., 2015; Kawa et al., 2018). One of these, the Multi-functional Fiber Laser Lidar (MFLL) (Dobbs et al., 2008; Dobler et al., 2013), has been flown (Campbell et al., 2020) as part of NASA's Atmospheric Carbon and Transport (ACT) – America project, an effort to detail CO2 variability as a function of weather and front location across the eastern half of North America (Davis et al., 2021). Several of these MFLL flights were designed to pass underneath OCO-2 along its ground track as it passed by, allowing CO2 from the better part of the full column to be compared between the two. Bell et al. (2020) have lined up the MFLL and OCO-2 data for these flights as a function of horizontal location and have assessed the accuracy of the along-track change in CO2 (the linear slope) from OCO-2 using the MFLL data as a ground truth: they accounted for the differences in the vertical averaging kernels and vertical extent between the two measurements in doing this comparison. Here, we use this same Bell et al. (2020) dataset to assess the along-track correlation length scale of the MFLL–OCO-2 measured column CO2 differences.
A team of inverse modelers using the OCO-2 retrievals have formed an OCO-2 flux inversion model intercomparison project (MIP) to help differentiate the CO2 fluxes robustly constrained by the OCO-2 data from the confounding biases in those same data (Crowell et al., 2019; Peiro et al., 2021): doing this as a MIP helps mitigate the impact of the errors in individual transport models (by looking at the results across the full ensemble of models) and the impact of any differences in the measurements and measurement errors used in the inversions (all MIP participants were directed to use the same 10 s average OCO-2 measurements and uncertainties). Here, we use the OCO-2 10 s averaging problem as an application to test the impact of the newly computed length scale. We interpret the MFLL–OCO-2 differences as OCO-2 retrieval errors and use the associated error correlation length scale to formulate a new model for the 10 s average value and its uncertainty. We then compare the results of this new error model to the results of the current error model to assess the impact of the newly calculated correlation length scale on both the absolute values of and the uncertainty calculated for the 10 s average values.
To better present the logical flow of our argument, we will split the paper into two parts, presenting both the method and results of our MFLL–OCO-2 analysis in the first (Sect. 2), then the method and results for the OCO-2 averaging application in the second. Since a few different averaging approaches have been used over time with the OCO-2 data, these previous methods will be outlined in this second section for context, before describing a new averaging approach using the correlation length scale. Section 3.1 presents a general framework for the OCO-2 data averaging approach, leaving the form of the correlation matrix, C, general. Section 3.1.1 discusses averages in which all the averaged retrievals are assumed to be independent (i.e., a diagonal C), Sect. 3.1.2 the case in which the errors in all the averaged retrievals are assumed to be correlated with all the others in the span with the same positive correlation coefficient (all the off-diagonal terms in C having the same value), and Sect. 3.1.3 the case in which the correlations are assumed to die out exponentially with distance along the satellite track (exponential decay off main diagonal in C). When averaging data with unequal uncertainty values, considering correlated errors can cause the average value computed to fall outside the range of the input values to be averaged. While this is a correct and natural consequence of the correlated error assumption, it does violate a key condition usually specified when defining a weighted average to prevent just that behavior: that all the weights be non-negative. Section 3.2 discusses this issue in more detail and lays out a couple of fallback options that we have used to stay with non-negative-weighted averages, while still reaping the benefits of the correlated error assumption. Section 3.3 applies these different correlation models (including both those with the fallback weighting and those without it) to two simple example cases, calculating the total measurement information content given by each model. Section 3.4 shows how the exponentially decaying correlation model (or “exponential” model hereafter) may be used to compute the effect of correlations between the full-span averages themselves (instead of the values going into them). In Sect. 3.5, we assess the impact of the correlation length scale determined from the MFLL–OCO-2 data on the averages of actual OCO-2 version 10 retrievals by calculating 10 s average values and their uncertainties using the new exponential correlation error model and comparing them to those given by the constant correlation error model. Finally, we discuss the implications of the new length-scale-dependent correlations in the conclusion.
2.1 MFLL measurements and their pairing with OCO-2 overflight data
The NASA-funded Atmospheric Carbon and Transport (ACT) – America project flew five aircraft campaigns over the 2016–2019 time period, measuring CO2, CH4, and meteorological variables in an effort to understand the relationship between atmospheric carbon and weather patterns, as well as the processes driving the uptake and release of carbon. These campaigns were done across all four seasons, each with flights in the mid-Atlantic, Midwest, and Gulf Coast regions of North America (Davis et al., 2021). As part of this effort, flights were made along the ground tracks of the OCO-2 satellite, instrumented with downward-viewing lidars taking CO2 measurements that could be compared with the column-averaged CO2 data taken by OCO-2.
The OCO-2 satellite measures radiances that are sensitive to dry air CO2 mixing ratios throughout the depth of the atmospheric column. CO2 mixing ratios are retrieved from these data at 20 levels in the vertical, evenly spaced in terms of pressure. Because there is not enough information to robustly differentiate these 20 values, a single pressure-weighted vertical average value, , is computed from these. How much information is contributed to the value from each level, as opposed to being taken from some prior guess, is shown by the shape of the vertical averaging kernel vector (see Fig. 6 of Bell et al., 2020): a value of 1 meaning all from the measurement and of 0 meaning all from the prior. The satellite flies in a sun-synchronous orbit, passing over the Equator at about 13:30 local time, as the Earth rotates underneath it; its 81∘ orbital inclination means that the flight path over North America is tilted somewhat, with the basic south-to-north motion going a bit SE to NW. Its 7077.7 km orbital semi-major axis results in a velocity of ∼6.75 km s−1 for its field of view (FOV) on the surface and gives 15 or 16 orbits per day. The observed path is at most 10 km wide, located along the ground track in nadir-viewing mode, and roughly parallel to it in glint mode: there are eight FOVs in the cross-track direction, each at most 1.25 km wide, and three cross-track scans are taken per second, making each FOV extend ∼2.25 km in the along-track direction.
The MFLL lidar (Dobbs et al., 2008; Dobler et al., 2013) was one of several flight instruments developed as a test bed for the CO2 lidar to be used aboard NASA's proposed ASCENDS satellite (Jucks et al., 2015; Kawa et al., 2018). We examine MFLL data from six OCO-2 underflights here, four taken over the Great Plains and two over the Mid-Atlantic (Table 1). The flight legs were generally around 500 km in length, with the aircraft taking about an hour to fly that distance. The satellite FOV, on the other hand, would take only about 75 s to traverse the same route so that, although the aircraft and satellite would be looking at very close to the same point on the ground at some point during the flight, the time difference in viewing could be up to 40 min or so at the ends of each leg: some change in the CO2 actually measured at the same location could thus be expected due to the blowing winds. The lidar was carried on a C-130 aircraft flying generally 8–9 km above ground level, or at about 350 hPa; because the OCO-2 measurements give lower weight to the upper parts of the column, the MFLL data are therefore able to provide an independent validation constraint on at least the lower of the OCO-2 column averages. The vertical weighting of the two measurements, as embodied in their averaging kernel vectors, is also different, with the MFLL instrument giving more weight to the upper part of the measured column just underneath the flight level, while the OCO-2 weight is more flat with pressure (see Fig. 6 of Bell et al., 2020). Lining up the two sets of measurements and accounting for the differences in vertical weighting has fortunately already been done by Bell (2018) and Bell et al. (2020). They used these data to test the accuracy of the spatial trend in CO2 retrieved by OCO-2 across the flight legs; here, we will use a subset of that same dataset to look at shorter-scale spatial variability (Baker et al., 2020). The reader is referred to Bell et al. (2020) for further details of the MFLL and OCO-2 measurements, as well as the comparison method and the details of the measurements on each flight leg.
2.2 Method for analyzing a correlation length scale
As described in Bell (2018) and Bell et al. (2020), the MFLL data have been binned and averaged across 60 s blocks corresponding to swaths of from 7 to 9 km in length along the OCO-2 ground track, depending on how fast the C-130 aircraft was flying at the time. All valid cloud-free OCO-2 retrievals (those declared “good” by the OCO-2 quality screening criteria) falling within the MFLL horizontal location range are similarly averaged. (Note, again, that the co-located MFLL and OCO-2 data may have measurement times that differ by up to an hour or more.) Only spans with more than about 20 MFLL and 3 OCO-2 measurements are used in the analysis. Here we use the satellite FOV latitude and longitude to calculate the distance between different measurement blocks. We subtract the OCO-2 average from its corresponding MFLL average for each common bin, then divide the difference Xj by the following OCO-2 measurement uncertainty value:
where the uncertainty from the retrieval, σj,retr, has been increased by a floor of 0.3 ppm to account for nonlinearities in the uncertainty calculation; this gives . We then detrend this weighted-difference time series across each flight leg – subtracting either a single constant value or a linear trend , where x is the along-track distance. The autocovariance function of this difference time series was then computed as
where Nh is the number of data values falling into each distance lag bin h (using an 8 km resolution) across all six flight legs. Finally, the autocorrelation coefficient values were computed as to give the autocorrelation spectrum shown in Fig. 1. A variety of sensitivity tests were performed, as well: (1) the correlations were computed without first dividing by the OCO-2 measurement uncertainties; (2) the data used in the analysis were restricted to pairs for which the MFLL and OCO-2 sampling times were within some threshold (3000, 1500, 1000 s); (3) whole flight legs of data were left out of the computation in succession; and (4) the pre-conditioning of the time series was switched between subtracting a constant value versus a linear trend. The only test that significantly changed the character of the spectrum was the last one.
2.3 Correlation length scale results
As shown in Fig. 1, the autocorrelation of the MFLL–OCO-2 differences falls off quickly, with no significant correlations at scales of more than about 20 km along-track. Even correlations at length scales shorter than 20 km are only weakly significant (at about the 1.5σ level); this analysis is pushing the boundaries of what this small set of comparison data can tell us. Still, for the case in which only a constant offset between the MFLL and OCO-2 time series is subtracted off, an exponentially decaying curve with a correlation length of 20 km does a reasonable job of fitting the spectrum, although there is a tendency for the actual spectrum to fall immediately to a correlation level of about +0.4 and then plateau somewhat out to a lag of 40 km. When the difference time series are detrended with a sloping line, the correlations drop off more quickly, with a correlation length of about 15 km, which is not surprising since more broad-scale information is removed by subtracting off the trend. We feel that subtracting off a sloped trend from spans of OCO-2 data this short (<600 km) is not appropriate, since the actual OCO-2 bias correction procedure is done globally and certainly leaves uncorrected gradients at these scales that ought to be considered in the analysis, so we believe that the longer length scale (20 km) would be more appropriate to apply as an OCO-2 error correlation length. (However, when averaging over very short spatial scales, say 10 km or less, the data would then support a correlation length scale of about 10 km due to the rapid initial fall.)
The magnitude of the correlated variability is given by multiplying the square roots of the correlation coefficients shown in Fig. 1 by the normalization factors obtained in computing them (0.59 ppm or 1.003 ppm when the trend is removed or not removed; or in terms of multiples of the uncertainty assumed on each MFLL–OCO-2 difference: 0.84 or 1.04σ). Moving over to 20 km on the x axis, a coefficient of 0.25 translates into a magnitude of 0.5× (0.59 or 1.003) = 0.3 or 0.5 ppm. This is large enough to be a significant fraction of the systematic errors in the OCO-2 data taken over land, which have been calculated to be about 0.6 ppm by Kulawik et al. (2019).
Because the MFLL data have been blocked into bins of 7 to 9 km in length, this analysis cannot resolve scales finer than that. That we obtain a length scale 2 to 3 times this minimum scale does suggest that it is real and not an artifact of the analysis. This 20 km length scale provides a significant constraint on the information content of the OCO-2 data in an average: it the scale was shorter, the uncertainty on longer averages would drop considerably, so if we consider the 20 km scale as an upper bound, we are being conservative (i.e., giving the data less weight in an inversion by applying a larger uncertainty to them). An exponential decay model of the shape of the correlations is not perfect but does not seem unreasonable to try when implementing the correlation structure found here.
We examine how the error correlation length scale derived above can help in weighting the column CO2 data from the OCO-2 satellite when used in global flux inversions. We focus here on the specific case of averaging the data across 10 s measurement spans (equivalent to an along-track distance of ∼67.5 km), but in the process we will get insight into how to handle data assimilated at both coarser scales and finer scales down to the 7–9 km MFLL binning size used here. We treat the MFLL data as the “truth” and interpret the entire MFLL–OCO-2 difference as an error in the retrieved OCO-2 values. An averaging approach that considers these correlation lengths is derived below, along with alternatives that use simpler assumptions, to help illustrate the differences caused by using our new approach.
3.1 Data averaging approach
Most estimation methods used in global atmospheric trace gas inversion work (Bayesian synthesis inversions, Kalman filters, variational data assimilation) combine measurement information in different time spans as
where xi is a vector of measurements within time span i and Ri the measurement error covariance matrix for xi. If R−1 is thought of as measurement “information” (i.e., the Fisher information matrix: see Rodgers, 2000), then represents the sum of the information in time spans 1 and 2, and xΣ can be thought of as a weighted average of x1 and x2 that summarizes their information. Note, however, that this approach veers wildly between extremes in its treatment of error correlations in time: measurement vectors x1 and x2 for different time spans are assumed to have errors that are uncorrelated with each other, while the elements of a measurement vector (at possibly different times inside the same time span) are permitted to have any nonzero correlations, as long as they may be described by an error covariance matrix. This assumption of uncorrelated errors between different time spans is built into the derivations of these inverse methods explicitly, for example in the Kalman filter, in which the dynamical errors related to propagating the measurement information from time to time are assumed to be uncorrelated with the measurement errors themselves (see Eq. 7.2-3 in Catlin, 1989, and Eq. 4.2-11 in Gelb, 1974). The more general case, in which errors between measurements at different times are considered to be correlated, may be written out and solved (see, e.g., Bennett, 2002, Sect. 1.5.3), but the greater computational complexity and workload involved, coupled with a general lack of knowledge as to what the temporal correlations ought to be assumed to be, result in the more simple forms being used more generally.
This same philosophy of considering the error correlations between measurements within a given time span, but neglecting them between time spans when assimilating them into inversions, has often been followed in assimilating OCO-2 CO2 data in global flux inversions: the measurements are first averaged across a certain time span (with any error correlations across these finer timescales and space scales being considered in the averaging process), then when these measurement averages are assimilated into the inversion, their errors are considered to be independent in a manner similar to that used in the inversion methods themselves.
The OCO-2 satellite makes three cross-scans per second, each of which spans only 10 km across-track and is divided into eight separate fields of view. Because the satellite “pirouettes” to keep the sun perpendicular to the viewing slit, this cross-scan is not always perpendicular to the ground track but can come within about 20∘ of being parallel to it (Eldering et al., 2017). Thus, OCO-2 senses only a very thin swath, up to 10 km wide, but which may be as thin as only 2 or 3 km at near-sub-solar latitudes. The satellite's FOV moves at ∼6.75 km s−1 along-track. Across a single second, then, OCO-2 takes up to 24 measurements across a quadrilateral with sides of 10 and 6.75 km, flattened to different degrees around the orbit. Not all of these 24 FOVs produce reliable retrievals due to clouds, high aerosol optical depths, or other problems that prevent the scene from passing the quality filters (xco2_quality_flag = 0 indicating a “good” scene).
Suppose we want to average the OCO-2 measurements across some distance along-track that will be closer to the grid box size of the typical atmospheric transport model used in global flux inversion studies (hundreds of kilometers on a side). The OCO-2 flux inversion MIP has averaged across a 10 s (67.5 km) swath, so we will use that here. If we form a vector of J (out of 240) “good” retrieved values to be averaged in our 10 s span, then their weighted average is calculated as
with 1 being a vector of ones. Choosing weights gives the scalar version of the sort of measurement information-weighted average shown in Eq. (3), which allows correlations between measurement errors within a given measurement span to be considered by specifying nonzero elements on the off-diagonal portion of the measurement error covariance matrix R. Suppose we break out the error covariance R into a form that explicitly represents the correlations as R=SCS, where C is the correlation matrix, with ones on the main diagonal and the correlation coefficients between the elements of measurement error dx on the off-diagonals, and where . The general form for the average then becomes
where and with the uncertainty in found from
Whether we want to consider correlations between these averages inside a given orbit is a separate matter. We might be forgiven from considering them to be independent in light of the choices made in the estimation methods themselves, especially if the along-track averaging length were to be long compared to the dominant error correlation length scale. But more on that later (in Sect. 3.4).
In the following subsections, we consider three different averages defined by Eq. (6), each using a different form for the correlation matrix C.
3.1.1 Averaging assuming uncorrelated errors
If the data values going into the average are assumed to have independent errors, setting the correlations to zero (C=I) gives
The summations here and hereafter are taken over , unless otherwise indicated, to simplify the notation. If σj=σo for all j, this gives the well-known result that .
One might use the straight information-weighted average given by Eq. (8) even in cases in which the errors are known to be correlated but when no good model for those correlations is available. In such a case, one could calculate an average uncertainty on the mean as follows:
This gives an average uncertainty on the mean that is similar in magnitude to the uncertainties of the averaged values rather than an uncertainty that decreases to reflect the sum of the incoming information. This former approach was the one used in the first attempt to compute 10 s averages from the OCO-2 data (using the version 7 release) – see Crowell et al. (2019) for details.
3.1.2 Averaging assuming constant correlations not depending on distance
Suppose that the error on each retrieval inside the averaging span is correlated with the error in every other retrieval inside the span with the same correlation coefficient, c, such that we have the J×J correlation matrix.
If we define H to be a J×J matrix with ones in every element, then , and (noting that H2=JH) we get
Based on work done by Susan Kulawik examining correlations between the OCO-2 retrievals, TCCON, and in situ aircraft measurements (personal communication, mid-March 2019), we used the following positive OCO-2 measurement correlation values in the MIP.
3.1.3 Averaging assuming correlations that decay exponentially with distance
If an error correlation length, L, is known, consider a 1-D error correlation model with positive correlations of the form . We will use the correlation length scale calculated for OCO-2 from the MFLL measurements for L. If we have J points to be averaged, spaced equally in the along-track direction and separated by distance Δx, then we have
with C−1 having the following convenient tridiagonal form:
To use this 1-D error correlation model for actual OCO-2 data, which may fall as much as 5 km on either side of the center of the ground track, some averaging in the cross-track direction must be done first. Once that is done, the data could then be averaged in the along-track direction at scales of anywhere from Δx=2.25 km (given by the 3 Hz cross-track scan frequency) all the way up to the 67.5 km distance traveled across the 10 s averaging span. To use the 1-D model with real data, for which there are data gaps, the missing data would be given sj values equal to zero in the formulas above.
The local nature of the average in Eq. (28) suggests a second use of this model: as a pre-conditioner for data to be assimilated retrieval by retrieval (individually, without averaging) in an inversion, with each measurement assumed to have errors independent of all the others. Currently, many inversion schemes ingest the data without averaging, retrieval by individual retrieval, sometimes inflating the uncertainties on individual data in an ad hoc manner to account for correlations and sometimes not. Equations (18) and (23) present a way to adjust the data beforehand so that when they are assimilated retrieval by retrieval, the correlations are accounted for in a statistically justifiable manner: pass over the data once before assimilating them, modifying each datum xj and its associated σj using the two data points on either side of it, along-track, such that
where the primes indicate the new adjusted values. Rather than being an approximation, this will give the same answer, when each datum is assimilated independently, that assimilating the original data, with correlations handled properly in the equations, would give.
3.2 Negative weights and their implications
In the definition of a weighted average, the weights on the averaged values are usually required to be non-negative, with at least one weight being positive. Non-negative weights can cause the averaged value to fall outside the range of the values to be averaged, a result that is generally considered undesirable in an average: the added requirement on the sign of the weights prevents this. However, under certain conditions the average values given by Eqs. (13) and (28) can give negative weights in Eq. (5) and out-of-range average values.
This out-of-range behavior was discovered when the constant correlation model from Sect. 3.1.2 was applied to the OCO-2 v9 data by the OCO-2 flux inversion MIP team. The weights R−11 for that model are
any element of which can go negative when or
Thus, most uncertainties that are larger than average (an average in terms of the inverse of the uncertainty) can cause negative weights.
The exponential correlation model can also yield negative weights. Taking an interior element of Eq. (18) as a guide, we have
which goes negative when
or (since ) when
This condition is violated only somewhat less frequently than Eq. (33): whenever σj is times greater than a similar average (of the inverses) of the uncertainties of the two neighboring points. When averaging many points at finer scales (Δx<L), , and up to half the points will cause negative weights.
A simple example helps explain why the error correlation models drive the weights negative and the average value out of range of the input values. Consider two data points, each measuring a quantity for which the true value is Xtrue=0: let the value and uncertainty on these points be x1=1, , . The error covariance matrix, given by
describes the (correlated) errors: the differences between the measurements and the truth, which, since the truth equals zero, are just the values of the measurements themselves. Then
Plugging this into Eq. (6) gives the correlated mean.
For β=1 (both uncertainties being the same), for all values of the correlation coefficient, c, except . But for β>1, moves more positive, closer to the measurement with more information or lower uncertainty, until , at which point . For larger β values, , which is outside the range of the two data values being averaged. Apparently, the correlated average, taking a clue from the value of the higher-uncertainty input, x2, believes that the errors on both x2 and (because of the positive correlation) x1 are negative (x2 being negative) and corrects for these errors by choosing a more positive value to the average value than the relative weighting of the two measurements, if uncorrelated, would otherwise require. When the difference between the uncertainties, β, is large enough (compared to ), the average value is driven outside the range of the input values. The weight on x2 is driven negative to achieve this. For this error correlation model, all this makes sense.
If the correlated averages given by Eqs. (13) and (28) are physically realistic, why not use them, even if they do not conform to the usual requirements of the weighted average? If it is clear that one's chosen model for the error correlations is correct, then yes, they should be used. But if one is not entirely sure of the model, that might be one reason to be hesitant to accept an average value that falls outside the range of the inputs. Is there an intermediate approach to fall back on that enforces the usual non-negative weight constraint for the average while still garnering the benefits of the correlated error models? One could try discarding the retrievals with higher retrieval uncertainties σj that seem to be driving the weights negative.
In the case of the constant correlation model, this would be impractical, as roughly half of the retrievals would have above-average σj values, and if those were thrown out, half of the remainder would have to be thrown out, and so on. For the exponential correlation model, however, a sort of filtering approach might be feasible when is significantly above 1 (i.e., for Δx values approaching L): one could throw out the retrievals with anomalously high σj values to the point that σj varied smoothly enough from retrieval to retrieval to ensure that the condition in Eq. (36) would never be violated. This might be practicable for spans with sparser data (longer Δx values) or when averaging together data that had already been binned together at finer scales.
As an alternative to using the negative weight criterion as a (possibly harsh) data filter, one might specify the form of the weighted average a priori in a way that forces the weights to be positive rather than letting the weights be determined indirectly by specifying the correlation model as we have done above. For example, one could specify the form of the average to be the information-weighted mean given in Eq. (8), but then impose the correlated error assumptions of one's choice when calculating the uncertainty on that mean. Since the mean given by Eq. (8) would, in general, no longer be the optimal (minimum variance) value for that error model, one might expect that the uncertainties on the mean obtained would be higher than those given by Eq. (14) or Eq. (22).
In general, if Cw is the correlation model assumed in setting the weights for the average and Cerr the correlation model assumed for the actual errors dx, then
3.2.1 Constant error correlation case with sub-optimal average
When the OCO-2 flux inversion MIP group encountered this out-of-range negative weight problem when applying Eq. (13) to the OCO-2 v9 data, this was in fact the work-around that we fell back to: the mean was specified by Eq. (8), but the errors between individual retrievals were assumed to be correlated according to the constant error correlation model from Eq. (11) (Peiro et al., 2021). Setting Cw=I and allows the uncertainty on the mean to be computed as
In terms of the measurement information, this gives
3.2.2 Exponentially decaying error correlation case with sub-optimal average
If one falls back to using Eq. (8) to calculate the means but still uses the exponential error correlation model, then (setting Cerr=C from Eq. 16) the uncertainty on the mean is computed (with ) as follows.
Or, in terms of measurement information,
3.3 Comparison of the error models for two simple cases
We look now at the uncertainty estimates that the error models above give for two simple example cases (Tables 2 and 3): one in which all σj=σo, a constant value, and a second in which varies across the sample as . Both cases may be solved analytically.
In Fig. 2, we plot for each of these correlation models as a function of the data spacing Δx for the two simple example cases. For the exponential correlation model, where , we divide the 10 s swath into equal increments . For the other error models, which use the number of averaged values J rather than Δx, we calculate for each point on the x axis, an assumption that forces all the data to be equally spaced along-track for those models, too, in these plots. The average uncertainty (dashed black) and independent error (green) models bound the possible information range as a function of Δx, with the independent case setting the maximum. There is more variation with J for the variable uncertainty case: for both simple models, there is a tendency for the constant correlation model to provide more rapidly increasing measurement information at smaller Δx than in the exponential correlation model: this is to be expected, since the correlations in the latter when Δx<20 km are higher than the +0.3 correlation level used in the constant correlation model. In general, the constant and exponential correlation models provide similar results, other than at the smallest Δx. The total information content of the average is limited by the correlations to only about 3 times that of the typical individual measurement when assuming an error correlation length scale of 20 km.
The results for the fallback approaches for the constant and exponential correlation models that we derived in Sect. 3.2.1 and 3.2.2 to get around the negative weight issue (in which we specified the weighted averaged to be given by Eq. 8 then computed its uncertainty using the correlated errors) are also shown in Fig. 2. For the exponential correlation model, the fallback approach (magenta lines) results in only slightly less information (or higher uncertainty) for the average compared to the original approach (red lines). For the constant correlation fallback model (cyan lines), the loss of information is greater, though this is seen only in the variable-σ case, with the two models giving the same result in the constant-σ case.
3.4 Calculating correlations between averaging spans
Above we have accounted for correlated errors between the retrievals going into the 10 s averages. The same 1-D error correlation model developed above can also be used to compute the correlations between adjacent 10 s average spans if an error correlation length scale is known. In that case, the spacing Δx between the data assumed in Sect. 3.1.3 and 3.2.2 is no longer the spacing between individual retrievals but rather the spacing between the different 10 s average spans along the ground track of the satellite.
Suppose we look at the daylit side of a single OCO-2 orbit, which we will assume encompasses a third of the full orbit, or about 13 358 km; there are about 198 10 s spans inside it. If we let , then we may use Eq. (22) or (45) again to compute the total information across the orbit. Figure 3 plots this information content for the two simplified error cases used above as a function of Δx: for our 10 s averages, we look at Δx=67.5 km, finding that the total information multiple is about 200 for the uncorrelated error case (green line), which is close to the J=198 value, as it should be. By comparing the information values for the uncorrelated error case (green line) to the values from the exponential correlation cases (red and magenta lines), we can see the impact of the correlations in reducing the total information content in the lower Δx range. Curves for three different correlation lengths (10, 20, and 40 km) are given; curves for measurement errors for the two different formulations of the mean (from Eqs. 22 and 45, the red and magenta curves in Fig. 2) are both plotted but fall on top of each other in this Δx range. By comparing the information assuming no correlations (green) versus exponential correlations (magenta), one can derive a single scalar multiple (greater than 1) of the uncertainties on the 10 s averages that will permit the 10 s averages to be assimilated in an inversion scheme with the assumption that they are all independent, but this will yield the same amount of total measurement information entering the problem as would have been obtained if the correlations had been accounted for properly using the original uncertainties. Using such an inflation factor is generally much easier than implementing the machinery for accounting for the correlations properly in the inversion code.
3.5 Application of the error correlation models to OCO-2 v10 data
To apply the exponential error correlation models presented in Sect. 3.1.3 and 3.2.2, one must somehow account for the fact that the real OCO-2 data are not one-dimensional but fall up to 5 km on either side of the center of the OCO-2 FOV ground track. As Fig. 2 shows that there is little independent information obtained from data spaced at Δx values much below the correlation length scale L, one would be justified in computing an average value at scales that fine using the independent error Eq. (8) and then applying a constant-correlation-based uncertainty to it using Eq. (43). To match the cross-track dimension (10 km) somewhat, we average all the OCO-2 v10 data values (Baker et al., 2020) falling inside each 2 s span (spaced Δx=13.5 km apart, along-track) in this manner. We could then apply any of the error correlation models that we have discussed so far (constant or exponential using the original or fallback forms for the weighted means) to the five 2 s average values falling inside each 10 s span to get the 10 s averages. As shown in Fig. 1, the MFLL–OCO-2 spectrum suggests that the error correlations fall off even more rapidly at small Δx than the 20 km correlation length that we decided was the best fit to the data across the full Δx range would suggest: we use a 10 km correlation length in calculating these 2 s averages to reflect that. An average Δx inside the 10 × 13.5 km averaging box would be about 6 km (keeping in mind the stretching of the box that occurs due to the pirouetting), giving an average correlation coefficient of . Our MFLL–OCO-2 data-based correlation estimates were taken over land and should apply only there: over the ocean, longer correlation lengths are generally assumed to apply. Since we do not have any MFLL data over the oceans as a guide, we will just use a correlation length double that over land there as a guess, giving .
Once the 2 s averages are computed, they then may reasonably be averaged up to 10 s values using our 1-D error model. This is done here for both the constant correlation and exponential correlation models using the same OCO-2 v10 data (Baker et al., 2020). Because of the difficulty satisfying the condition that all weights remain positive in the full constant correlation case presented in Sect. 3.1.2, we will not calculate results for that model but rather only for the fallback model presented in Sect. 3.2.1. We will compute this case using both the two-step process (averaging first across 2 s spans, then averaging those across 10 s spans) used for the exponential correlation models and a one-step process of averaging all retrievals inside the 10 s span in a single step, without applying the shorter correlation length scales inside the 2 s boxes. For the exponential correlation models, we will calculate results for both the original model (from Sect. 3.1.3) with the data screened to satisfy the positive weight criterion and the fallback model (from Sect. 3.2.2).
Since we consider the exponential correlation model to be an improvement to the constant correlation model (with the fallback weights) given in Sect. 3.2.1, we are interested in how much this new model shifts the average from the old one. We also compute the difference in the uncertainties for the four cases. We double the correlation length over the oceans from 20 to 40 km for the calculation of the 10 s averages as well.
All 2 s spans without any data have their values set to zero. Those 2 s σj values that produce negative weights according to Eq. (36) have their values set to zero as well to ensure that the average value stays within the range of the input values. The 2 s average values thrown out by this approach will tend to be those with higher uncertainties sandwiched in between adjacent spans with lower uncertainties; those next to a span with no data are less likely to be discarded. Over the ocean, where the correlation length is doubled, with , it could be expected that more of the 2 s averages will be discarded than over land, where . That was, in fact, the case. However, uncertainties for adjacent 2 s spans do not seem to vary too much: only about 2.5 % of the ocean scenes needed to be discarded and less than 1 % of them over land for the 2 s averaging spans. The 2 s span length was chosen over the 1 s one to avoid having to throw out more data than this: for the 1 s span, ∼10 % and ∼7 % of the ocean and land data would have had to be discarded, respectively.
Figure 4 shows the number of OCO-2 10 s averages per 2∘ × 2∘ bin across September 2014–October 2020, broken into two halves of the year (April–September and October–March), while Fig. 5 gives the 10 s average values for the same spans for the constant correlation model from Sect. 3.2.1 acting upon the 2 s means (i.e., using the two-step approach from above). Figure 6 then gives the difference in 10 s averaged from this two-step model for both the one-step constant correlation model and the exponential correlation model. It is interesting that the exponential model, which uses completely different assumptions about the correlations as a function of Δx, gives almost identical averages as the two-step constant correlation model. The difference to the constant correlation averages caused by using the intermediate step of averaging the data across a 2 s span is a larger effect. (Recall that we had to use a two-step process for the exponential correlation model to be able to apply the 1-D error model to it in the first place. So we compute results for a similar two-step constant correlation model as well to allow a more accurate comparison of the two approaches.) The close agreement between the exponential model average and the two-step constant model average is somewhat deceptive, though: when the same comparison (not shown) was done using 1 s averages for both, there was a systematic difference between the averages over the land and ocean of over 0.1 ppm with opposite sign. This shift may be related to the greater number of points being thrown out for causing negative weights in the 1 s averaging case.
Figure 7 presents the uncertainty of the 2 s averages going into the two-step constant correlation averages (calculated from the uncertainties for the individual retrievals using Eq. 42), while Fig. 8 gives the ratio of total measurement information given by four different 10 s correlation models with respect to these 2 s uncertainty values. These information multiples are the same as those plotted on the y axis of Fig. 2 for the two simple error cases and are meant to show the number of independent pieces of information allowed by the correlations across the 10 s span (a value between 1 and 5). Both of the models that addressed the negative weight issue by falling back to the independent-error average of Eq. (8), shown in the right column of Fig. 8, are similar and yield slightly more information (or lower uncertainties on the 10 s averages) than the one-step constant correlation model (upper left). The two-step constant correlation model gives somewhat more information over the oceans than the one-step model, despite imposing higher correlations during the 2 s averaging: greater weight given to 2 s spans with sparser data may explain the difference. The exponential correlation model (lower left) gives significantly higher measurement information content (lower uncertainty) than the other three correlation models, especially over the oceans. The large difference over the oceans is in agreement with what is seen in Fig. 2 for the simple model with variable σj (right panel), where the dotted orange line (exponential model for L=40 km) at Δx=13.5 km is about 25 % higher than the dashed cyan line (showing the constant correlation result for ). In contrast, the values over land are similar between the models in that figure: compare the solid orange and magenta lines (for the original and fallback exponential models, assuming L=20 km) to the solid cyan ( constant correlation) curve. The information loss in going from a correlation length of 20 km over land to 40 km over the ocean in the exponential model is not as large as the loss of information in going from a correlation of +0.3 to +0.6 in the constant correlation model, at least at Δx=13.5 km: that explains the different character of the results over the oceans versus land.
The exponential correlation model seems to be quite sensitive to data dropout: if every other 2 s average along-track is thrown out, the total information returned in the average goes up, as can be seen from Eq. (21). This is counterintuitive and is a feature of this model that deserves more investigation.
It has long been recognized in the atmospheric modeling community that “measurement errors” in flux inversions must include not just instrumentation errors, but also errors incurred when representing the measurements in the models, especially the coarse-resolution transport models used in global inversions. In the case of in situ measurements, the former might be on the order of 0.1 or 0.2 ppm, while the latter could range from as low as 0.2 ppm in the remote Southern Hemisphere to multiple parts per million (ppm) for continental sites farther north, especially those feeling the effects of nearby forests or cities. Correlations between measurements located near each other in space or time should similarly be due to modeling errors in these “model–data mismatch” (MDM) errors. In inversions using just in situ data, the MDM errors would be increased to de-weight multiple sites located close to each other (e.g., Bermuda East and West) or multiple data streams from different measurement groups or different types of sensors at a single site (Mauna Loa, Cape Grim, South Pole). To account for diurnal modeling errors, continuous measurements would be de-weighted versus daily measurements and both against weekly flask measurements. Aircraft profiles spanning multiple vertical levels might be de-weighted to account for vertical mixing errors.
For satellites such as GOSAT, which take discrete column-averaged measurements spaced generally over 100 km apart from each other around the orbit, the need for modeling correlations between measurements was less immediate: a modeling error could be added to the retrieval error in quadrature in a plausible error treatment. For a satellite like OCO-2, however, with up to 1000 measurements taken in a thin swath across a 300 km span, there are sure to be correlations in the retrieval errors on those scales (due to the parameters assumed in the retrievals, and the modeling errors on them, varying on those scales), not to mention atmospheric modeling errors as well.
For the OCO-2 version 7 data, the OCO-2 flux inversion MIP team used a two-step averaging approach: the data were first averaged across a 1 s span using an information average (Eq. 8), and then an uncertainty was placed upon this value that was a combination of the average uncertainty given by Eq. (10), derived from the retrieval uncertainties, and the standard deviation of the retrieved values going into the average. Those 1 s spans with good data were then averaged using Eq. (8) with an uncertainty placed upon the average that assumed that each 1 s span had errors independent of all the others, according to Eq. (9). Finally, a modeling error was added in quadrature to the MDM error so calculated: this modeling error ended up being so large compared to the measurement-derived error that, in practice, the details of the measurement-derived part did not matter much. See Crowell et al. (2019) for details.
For version 9 of the OCO-2 retrievals, the OCO-2 flux inversion MIP team attempted to do a better job modeling correlations between the values. Instead of neglecting them, the correlations were set to constant values of +0.3 over land and +0.6 over ocean (see Eq. 15), with no dependence on separation distance considered. To use these, the MIP team derived the constant correlation model outlined in Sect. 3.1.2 and attempted to apply it to the v9 data, but they discovered that it yielded an average value that often fell outside the range of the input values and determined that this was due to the weights on individual terms in the average being calculated to be negative by the model. Because the team was uncertain how physically realistic these average values were at that time, they fell back to using the model presented in Sect. 3.2.1 instead: the average was calculated using the old information average of Eq. (8), with a newly calculated uncertainty given by Eq. (42) using the constant correlation coefficients. The average was performed in a single step across the 10 s span, without first computing averages at shorter spans. (This was because Kulawik et al., 2019, had found that giving the sparser OCO-2 data more weight via the previous two-step approach resulted in a poorer fit to the TCCON data: the sparser data, while providing better spatial coverage, are also apparently more susceptible to errors due to nearby clouds or other scatterers.)
The MFLL–OCO-2 differences that we have analyzed here suggest that the assumption that the OCO-2 error correlations are constant across scales all the way up to the 67.5 km 10 s averaging length is not a good one: if the MFLL–OCO-2 differences are, in fact, a good proxy for errors on the full column1, then Fig. 1 shows that their correlations drop off rapidly, with a decorrelation length of about 20 km fitting the data well across most of the measured spectrum. These data also suggest that much of the correlation in the differences falls off even more rapidly to a coefficient of below +0.4 at the finest resolvable scale of 8 km, corresponding to a correlation length of more like 10 km at these finest scales. These two scales might be due to separate error sources at the two different Δx scales: for example, more quickly changing errors due to surface-related parameters in the retrieval (albedo, pressure over topography) versus more slowly changing errors due to atmosphere-related parameters such as water vapor, temperature, or aerosols.
With these correlation length scales in hand, we derived a one-dimensional error model here with exponentially decaying correlations as a function of the along-track distance (Sect. 3.1.3). We discovered that this model produced negative weights just as our constant correlation model did, though to a lesser extent, and derived a similar fallback model as was used with the constant correlation approach as one way to get around this problem (Sect. 3.2.2). We applied all these correlation models to two simple cases that could be solved analytically to get a better understanding of how the models behaved with different averaging assumptions (different separations Δx of the incoming average values).
Finally, we applied both our new exponential model and our old constant correlation model to the recently released OCO-2 version 10 data. For the exponential model, we first averaged the data across a 2 s span to force the cross-track retrievals onto the 1-D satellite FOV ground track. By choosing this pre-averaging span to be 2 s long, we were able to reduce the number of spans that had to be thrown out for violating the positive weight criterion to just a couple of percent, allowing the original exponential model (from Sect. 3.1.3) to be applied for a subset of the data similar to that used in the other error models. The new exponential model caused generally negligible shifts in the average values compared to the constant correlation model at seasonal to annual scales. This is good news for our previous results and suggests that the constant correlation coefficients that we used in the OCO-2 v9 MIP studies (Peiro et al., 2021) do a good job approximating the correlations across the 10 s spans on average, even though those appear to vary as a function of the distance. Shifts in for individual retrievals are found at the level of a few tenths of 1 ppm, and systematic shifts in seasonal and annual between land and ocean of from 0.1 to 0.2 ppm were found when pre-averaging across 1 s rather than 2 s spans.
Since the MFLL data were only taken over land, they cannot say what the OCO-2 correlation length scale might be over the oceans. As a guess, we have simply doubled the land values here to get ocean values. For that assumption, the exponential correlation model gives uncertainties on the 10 s averages that are significantly lower than those given by the constant correlation model using the +0.6 coefficient value. Possibly, this indicates that we should use longer correlation lengths over the ocean than simply doubling the land values. But from the differences seen between the constant and exponential correlation models at coarser scales in the right panel of Fig. 2, this may indicate that the new exponential correlation model is providing better uncertainty estimates (ones which, when used in the inversions, will give more weight to the OCO-2 data with respect to the prior information). Lidar underflight data similar to the MFLL data used here, but collected over the ocean, are much needed to help extend the utility of this approach globally.
The 1-D error model with exponentially decaying correlations that we have derived here also has immediate application in two other areas: (1) in the calculation of correlations between adjacent 10 s averaging spans and (2) in the treatment of correlated satellite data in inversion schemes that choose not to average the individual retrievals before assimilating them but choose to handle the correlations inside the scheme itself. Regarding point (1), by comparing the uncertainties given by the exponential correlation model (with Δx equal to the distance between the 10 s span, or whatever longer averaging span is being considered) to those given by assuming that the error in each 10 s span is independent from all others (i.e., by comparing the magenta and green curves from Fig. 3), an overall inflation factor may be computed, which, when the uncertainties on individual 10 s average values are multiplied by this factor, will allow them to be assimilated in the inversion assuming independent errors but will still give the proper weight vis-à-vis the prior information in the problem that would have been obtained had the correlations between 10 s averages been modeled explicitly in the inversion. Similarly, regarding point (2), Eqs. (30) and (29) provide a way to adjust the pre-averaged measurements (the 2 s averages in our presentation here) and their uncertainties in a manner such that when each of these measurements is assimilated in the inversion assuming independent errors, the same answer is obtained as if the correlations between the errors in each measurement were explicitly modeled in the inversion itself (e.g., by using off-diagonal terms in R). We make this argument using the 2 s averages instead of the original retrievals themselves because the original retrievals are not strictly laid out in a one-dimensional string; they have a distribution extending up to 10 km in the cross-track direction that must be dealt with before the 1-D model can be applied.
The one-dimensional model that we have examined here can provide some insight that might be useful for satellites whose data extend across wider swaths, like the proposed CarbonSat mission, or to missions that scan full continents in a truly two-dimensional manner (GeoCarb). As Fig. 2 shows, the correlation length scale sets a limit on the total number of independent pieces of information contained across a certain span: as a very rough approximation, the number of independent pieces of information is about equal to the total averaging span (e.g., 67.5 km) divided by the correlation length scale. In going to the more general 2-D problem, this suggests that the total information across a given area is proportional to the number of squares (of the correlation length wide) that will fit into the area. The uncertainties of individual shots within the area may then be scaled such that their total information, when they are assimilated independently, equals that given from the correlation length analysis.
The data and code used in this analysis may be downloaded from the CERN-based Zenodo archive at https://doi.org/10.5281/zenodo.4399884 (Baker et al., 2020). This includes co-located MFLL and OCO-2 column CO2 measurements for six underflights and the MATLAB script used to calculate the autocorrelation spectrum from their differences, as well as OCO-2 version 10 data and the Fortran programs used to calculate the 2 and 10 s averages from them.
JD and BL have designed, implemented, developed, and tested the MFLL instrument, as well as helped with its proper usage on the ACT flight campaigns. JFC helped with the usage of MFLL on the ACT campaigns and evaluated its performance. KJD designed and oversaw the ACT campaigns. EB obtained MFLL and OCO-2 data, co-located them in space and time, accounted for the differences in vertical averaging kernel between them, and produced files with them side by side. DFB computed the correlation spectrum of the MFLL–OCO-2 differences and developed and applied the error correlation models in the remainder of the document.
The contact author has declared that neither they nor their co-authors have any competing interests.
Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
David F. Baker would like to thank the OCO-2 flux inversion MIP team for many discussions regarding the use of the constant error correlation models presented in Sect. 3.1.2 and 3.2.1. In particular, he would like to acknowledge the help of the following: Sourish Basu, who co-derived the constant error model presented in Sect. 3.1.2, discovered that using it led to out-of-range average values when it was first applied to the OCO-2 v9 data, and pointed out that this was due to the weights in the average going negative; Susan Kulawik for her work in deriving the constant error correlation values in Eq. (15) and for encouraging the publication of the correlation length scale work presented in Sect. 2; and Jonathan Hobbs, Hai Nguyen, and Vineet Yadav for pointing out that one is always free to define the weights in an average as one wishes, then to apply an error model when computing the uncertainty in that defined average – this led to the error models presented in Sect. 3.2.1 and 3.2.2, the first of which was adopted by the OCO-2 flux inversion MIP team in the work using the OCO-2 v9 data. The authors would also like to thank the ACT-America, MFLL, and OCO-2 teams for collecting and processing the data used here and making it available to other researchers.
This research has been supported by the National Aeronautics and Space Administration (“Atmospheric Carbon and Transport study – America (ACT-America)”, grant no. NNX15AJ07G).
This paper was edited by Juan Antonio Añel and reviewed by two anonymous referees.
Baker, D. F., Bell, E., Davis, K. J., Campbell, J. F., Lin, B., and Dobler, J.: Computing a correlation length scale from MFLL–OCO-2 CO2 differences, and accounting for correlated errors when assimilating OCO-2 data, Zenodo [data set], https://doi.org/10.5281/zenodo.4399884, 2020. a, b, c, d
Bell, E.: Evaluation of OCO-2 Small-scale Variability Using Lidar Retrievals from the ACT-America Flight Campaign, MS thesis, Dept. of Atmospheric Science, Colorado State University, U.S.A., 114 pp., available at: https://mountainscholar.org/handle/10217/191457 (last access: 17 January 2022), 2018. a, b, c
Bell, E., O'Dell, C. W., Davis, K. J., Campbell, J., Browell, E., Denning, A. S., Dobler, J., Erxleben, W., Fan, T.-F., Kooi, S., Lin, B., Pal, S., and Weir, B.: Evaluation of OCO‐2 Variability at Local and Synoptic Scales using Lidar and In Situ Observations from the ACT‐America Campaigns, J. Geophys. Res.-Atmos., 125, e2019JD031400, https://doi.org/10.1029/2019JD031400, 2020. a, b, c, d, e, f, g, h
Bennett, A. F.: Inverse Modeling of the Ocean and Atmosphere, Cambridge University Press, New York, USA, ISBN 13 978-0-521-02157-9, ISBN-10 0-521-02157-X, 2002. a
Campbell, J. F., Lin, B., Dobler, J., Pal, S., Davis, K., Obland, M. D., Erxleben, W., McGregor, D., O'Dell, C., Bell, E., Weir, B., Fan, T.-F., Kooi, S., Gordon, I., Corbett, A., and Kochanov, R.: Field evaluation of column CO2 retrievals from intensity‐modulated continuous‐wave differential absorption lidar measurements during the ACT‐America campaign, Earth and Space Science, 7, e2019EA000847, https://doi.org/10.1029/2019EA000847, 2020. a
Catlin, D. E.: Estimation, Control, and the Discrete Kalman Filter, in: Applied mathematical sciences, volume 71, edited by: John, F., Marsden, J. E., and Sirovich, L., Springer-Verlag, New York, USA, ISBN 0-387-96777-X, ISBN 3-540-96777-X, 1989. a
Crisp, D., Atlas, R. M., Breon, F.-M., Brown, L. R., Burrows, J. P., Ciais, P., Connor, B. J., Doney, S. C., Fung, I. Y., Jacob, D. J., Miller, C. E., O’Brien, D., Pawson, S., Randerson, J. T., Rayner, P., Salawitch, R. J., Sander, S. P., Sen, B., Stephens, G. L., Tans, P. P., Toon, G. C., Wennberg, P. O., Wofsy, S. C., Yung, Y. L., Kuang, Z., Chudasama, B., Sprague, G., Weiss, B., Pollock, R., Kenyon, D., and Schroll, S.: The orbiting carbon observatory (OCO) mission, Adv. Space Res., 34, 700–709, https://doi.org/10.1016/j.asr.2003.08.062, 2004. a
Crisp, D., Miller, C. E., and DeCola, P. L.: NASA Orbiting Carbon Observatory: Measuring the column averaged carbon dioxide mole fraction from space, J. Appl. Remote Sens., 2, 023508, https://doi.org/10.1117/1.2898457, 2008. a
Crowell, S., Baker, D., Schuh, A., Basu, S., Jacobson, A. R., Chevallier, F., Liu, J., Deng, F., Feng, L., McKain, K., Chatterjee, A., Miller, J. B., Stephens, B. B., Eldering, A., Crisp, D., Schimel, D., Nassar, R., O'Dell, C. W., Oda, T., Sweeney, C., Palmer, P. I., and Jones, D. B. A.: The 2015–2016 carbon cycle as seen from OCO-2 and the global in situ network, Atmos. Chem. Phys., 19, 9797–9831, https://doi.org/10.5194/acp-19-9797-2019, 2019. a, b, c
Davis, K. J., Browell, E. V., Feng, S., Lauvaux, T., Obland, M. D., Pal, S., Baier, B. C., Baker, D. F., Baker, I. T., Barkley, Z .R., Bowman, K. W., Cui, Y. Y., Denning, A. S., Digangi, J. P., Dobler, J. T., Fried, A., Gerken, T., Keller, K., Ling, B., Nehrir, A. R., Normile, C. P., O’Dell, C. W., Ott, L. E., Roiger, A., Schuh, A. E., Sweeney, C., Wei, Y., Weir, B., Xue, M., and Williams, C. A.: The Atmospheric Carbon and Transport (ACT) – America NASA Earth Venture Suborbital Mission, B. Am. Meteorol. Soc., 102, E1714–E1734, https://doi.org/10.1175/BAMS-D-20-0300.1, 2021. a, b
Dobbs, M., Krabill, W., Cisewski, M., Harrison, F. W., Shum, C. K., McGregor, D., Neal, M., and Stokes, S.: A multi-functional fiber laser lidar for earth science and exploration, in: IGARSS 2008–2008 IEEE International Geoscience and Remote Sensing Symposium, vol. 3, Boston, Massachusetts, IEEE, III-350–III-353, https://doi.org/10.1109/IGARSS.2008.4779355, 2008. a, b
Dobler, J. T., Harrison, F. W., Browell, E., Lin, B., McGregor, D., Kooi, S., Choi, Y., and Ismail, S.: Atmospheric CO2 column measurements with an airborne intensity‐modulated continuous wave 157 µm fiber laser lidar, Appl. Optics, 52, 2874, https://doi.org/10.1364/AO.52.002874, 2013. a, b
Eldering, A., O'Dell, C. W., Wennberg, P. O., Crisp, D., Gunson, M. R., Viatte, C., Avis, C., Braverman, A., Castano, R., Chang, A., Chapsky, L., Cheng, C., Connor, B., Dang, L., Doran, G., Fisher, B., Frankenberg, C., Fu, D., Granat, R., Hobbs, J., Lee, R. A. M., Mandrake, L., McDuffie, J., Miller, C. E., Myers, V., Natraj, V., O'Brien, D., Osterman, G. B., Oyafuso, F., Payne, V. H., Pollock, H. R., Polonsky, I., Roehl, C. M., Rosenberg, R., Schwandner, F., Smyth, M., Tang, V., Taylor, T. E., To, C., Wunch, D., and Yoshimizu, J.: The Orbiting Carbon Observatory-2: first 18 months of science data products, Atmos. Meas. Tech., 10, 549–563, https://doi.org/10.5194/amt-10-549-2017, 2017. a, b
Gelb, A. (Ed.): Applied Optimal Estimation, The M.I.T. Press, Cambridge, Mass., USA, ISBN 0-262-57048-3, 1974. a
Jucks, K. W., Neeck, S., Abshire, J. B., Baker, D. F., Browell, E. V., Chatterjee, A., Crisp, D., Crowell, S. M., Denning, S., Hammerling, D., Harrison, F., Hyon, J. J., Kawa, S. R., Lin, B., Meadows, B. L., Menzies, R. T., Michalak, A., Moore, B., Murray, K. E., Ott, L. E., Rayner, P., Rodriguez, O. I., Schuh, A., Shiga, Y., Spiers, G. D., Wang, J. S., and Zaccheo, T. S.: Active Sensing of CO2 Emissions over Nights, Days, and Seasons (ASCENDS) Mission, Science Mission Definition Study, available at: https://cce.nasa.gov/ascends_2015/ASCENDS_FinalDraft_4_27_15.pdf (last access: 14 January 2022), 2015. a, b
Kawa, S. R., Abshire, J. B., Baker, D. F., Browell, E. V., Crisp, D., Crowell, S. M. R., Hyon, J. J., Jacob, J. C., Jucks, K. W., Lin, B., Menzies, R. T., Ott, L. E., and Zaccheo, T. S.: Active Sensing of CO2 Emissions over Nights, Days, and Seasons (ASCENDS): Final Report of the ASCENDS Ad Hoc Science Definition Team, NASA/TP-2018-219034, available at: https://www-air.larc.nasa.gov/missions/ascends/docs/NASA_TP_2018-219034_ASCENDS_ID1.pdf (last access: 17 January 2022), 2018. a, b
Kulawik, S. S., Crowell, S., Baker, D., Liu, J., McKain, K., Sweeney, C., Biraud, S. C., Wofsy, S., O'Dell, C. W., Wennberg, P. O., Wunch, D., Roehl, C. M., Deutscher, N. M., Kiel, M., Griffith, D. W. T., Velazco, V. A., Notholt, J., Warneke, T., Petri, C., De Mazière, M., Sha, M. K., Sussmann, R., Rettinger, M., Pollard, D. F., Morino, I., Uchino, O., Hase, F., Feist, D. G., Roche, S., Strong, K., Kivi, R., Iraci, L., Shiomi, K., Dubey, M. K., Sepulveda, E., Rodriguez, O. E. G., Té, Y., Jeseck, P., Heikkinen, P., Dlugokencky, E. J., Gunson, M. R., Eldering, A., Crisp, D., Fisher, B., and Osterman, G. B.: Characterization of OCO-2 and ACOS-GOSAT biases and errors for CO2 flux estimates, Atmos. Meas. Tech. Discuss. [preprint], https://doi.org/10.5194/amt-2019-257, 2019. a, b, c
O'Dell, C. W., Connor, B., Bösch, H., O'Brien, D., Frankenberg, C., Castano, R., Christi, M., Eldering, D., Fisher, B., Gunson, M., McDuffie, J., Miller, C. E., Natraj, V., Oyafuso, F., Polonsky, I., Smyth, M., Taylor, T., Toon, G. C., Wennberg, P. O., and Wunch, D.: The ACOS CO2 retrieval algorithm – Part 1: Description and validation against synthetic observations, Atmos. Meas. Tech., 5, 99–121, https://doi.org/10.5194/amt-5-99-2012, 2012. a
O'Dell, C. W., Eldering, A., Wennberg, P. O., Crisp, D., Gunson, M. R., Fisher, B., Frankenberg, C., Kiel, M., Lindqvist, H., Mandrake, L., Merrelli, A., Natraj, V., Nelson, R. R., Osterman, G. B., Payne, V. H., Taylor, T. E., Wunch, D., Drouin, B. J., Oyafuso, F., Chang, A., McDuffie, J., Smyth, M., Baker, D. F., Basu, S., Chevallier, F., Crowell, S. M. R., Feng, L., Palmer, P. I., Dubey, M., García, O. E., Griffith, D. W. T., Hase, F., Iraci, L. T., Kivi, R., Morino, I., Notholt, J., Ohyama, H., Petri, C., Roehl, C. M., Sha, M. K., Strong, K., Sussmann, R., Te, Y., Uchino, O., and Velazco, V. A.: Improved retrievals of carbon dioxide from Orbiting Carbon Observatory-2 with the version 8 ACOS algorithm, Atmos. Meas. Tech., 11, 6539–6576, https://doi.org/10.5194/amt-11-6539-2018, 2018. a
Peiro, H., Crowell, S., Schuh, A., Baker, D. F., O'Dell, C., Jacobson, A. R., Chevallier, F., Liu, J., Eldering, A., Crisp, D., Deng, F., Weir, B., Basu, S., Johnson, M. S., Philip, S., and Baker, I.: Four years of global carbon cycle observed from OCO-2 version 9 and in situ data, and comparison to OCO-2 v7, Atmos. Chem. Phys. Discuss. [preprint], https://doi.org/10.5194/acp-2021-373, in review, 2021. a, b, c
Rodgers, C. D.: Inverse Methods for Atmospheric Sounding: Theory and Practice, in: Series on Atmospheric, Oceanic, and Planetary Physics, vol. 2, edited by: Taylor, F. W., World Scientific Publishing Co., Hackensack, NJ, USA, ISBN 13 978-981-02-2740-1, ISBN 10 981-02-2740-X, 2000. a
Worden, J. R., Doran, G., Kulawik, S., Eldering, A., Crisp, D., Frankenberg, C., O'Dell, C., and Bowman, K.: Evaluation and attribution of OCO-2 XCO2 uncertainties, Atmos. Meas. Tech., 10, 2759–2771, https://doi.org/10.5194/amt-10-2759-2017, 2017. a
Wunch, D., Wennberg, P. O., Osterman, G., Fisher, B., Naylor, B., Roehl, C. M., O'Dell, C., Mandrake, L., Viatte, C., Kiel, M., Griffith, D. W. T., Deutscher, N. M., Velazco, V. A., Notholt, J., Warneke, T., Petri, C., De Maziere, M., Sha, M. K., Sussmann, R., Rettinger, M., Pollard, D., Robinson, J., Morino, I., Uchino, O., Hase, F., Blumenstock, T., Feist, D. G., Arnold, S. G., Strong, K., Mendonca, J., Kivi, R., Heikkinen, P., Iraci, L., Podolske, J., Hillyard, P. W., Kawakami, S., Dubey, M. K., Parker, H. A., Sepulveda, E., García, O. E., Te, Y., Jeseck, P., Gunson, M. R., Crisp, D., and Eldering, A.: Comparisons of the Orbiting Carbon Observatory-2 (OCO-2) measurements with TCCON, Atmos. Meas. Tech., 10, 2209–2238, https://doi.org/10.5194/amt-10-2209-2017, 2017. a
There are good reasons why the MFLL–OCO-2 difference might not be a good proxy for errors in OCO-2 retrievals: much of the difference might be due to errors in the MFLL retrieval, to problems matching the two measurement types in time and space, or to errors in accounting for the different vertical averaging kernels; also, only the lower of the OCO-2 column may be compared to the MFLL data given the cruise height of the aircraft.