Blockworlds 0.1.0: a demonstration of anti-aliased geophysics for probabilistic inversions of implicit and kinematic geological models

. Parametric geological models such as implicit or kinematic models provide low-dimensional, interpretable representations of 3-D geological structures. Combining these models with geophysical data in a probabilistic joint inversion framework provides an opportunity to directly quantify uncertainty in geological interpretations. For best results, care must be taken with the intermediate step of ren-dering parametric geology in a ﬁnite-resolution discrete basis for the geophysical calculation. Calculating geophysics from naively voxelized geology, as exported from commonly used geological modeling tools, can produce a poor approximation to the true likelihood, degrading posterior inference for structural parameters. We develop a simple integrated Bayesian inversion code, called Blockworlds, showcasing a numerical scheme to calculate anti-aliased rock properties over regular meshes for use with gravity and magnetic sensors. We use Blockworlds to demonstrate anti-aliasing in the context of an implicit model with kinematic action for simple tectonic histories, showing its impact on the structure of the likelihood for gravity anomaly.


Introduction
Geological modeling of subsurface structures is critical to decision-making across numerous application areas, including mining, groundwater, resource exploration, natural hazard assessment, and engineering, yet is also subject to considerable uncertainty (e.g., Pirot et al., 2015;Linde et al., 2017;Witter et al., 2019;Quigley et al., 2019;Wang et al., 2020). Uncertainty arises in numerous places within the modelbuilding workflow, including sparse, noisy, and/or heterogeneous geological observations and rock property measurements; the indirect nature of geophysical measurements; the non-uniqueness of inverse problem solutions; and the ambiguity of geological interpretations (Lindsay et al., 2013). Rigorous quantification of uncertainty is therefore critical to decision-making informed by models, and is becoming an increasingly active area of research in geology and geophysics (Rawlinson et al., 2014;Bond, 2015;Jessell et al., 2018;Wellmann and Caumon, 2018).
Bayesian methods provide a probabilistically coherent framework for reasoning about uncertainty (Tarantola and Valette, 1982;Mosegaard and Tarantola, 1995;Mosegaard and Sambridge, 2002). Although not the only way to account for uncertainty in earth science settings, Bayesian reasoning enables the natural integration of heterogeneous data and expert knowledge (Jessell et al., 2010;Bosch, 2016;Beardsmore et al., 2016), guides the acquisition of additional data for maximum information gain (e.g., Pirot et al., 2019b), enables selection among competing conceptual models (e.g., Brunetti et al., 2019;Pirot et al., 2019a), and optimizes management of risk in decision-making over possible outcomes (e.g., Varouchakis et al., 2019).
Bayesian methods are used in both geology and geophysics, but with different framing for inference problems. In the geophysics community, where quantitative inversion frameworks have been in use for decades (Backus and Gilbert, 1967, 1968, 1970, the core problem is framed as a nonparametric imaging of the subsurface. The model parameters directly define discretized spatial fields of rock properties, from which forward geophysics are calculated to reproduce observations. Model structures rich enough to be useful are often very high-dimensional, and variations in spatial scales and resolutions make the full covariance matrix cumbersome and impractical to evaluate (Rawlinson et al., 2014). Although the discrete geophysical image representation may naturally reflect the geological interpretation (for example, in the transdimensional inversions reviewed in Sambridge et al., 2012), this is not usually the case, and much effort goes into developing models with priors or regularizing terms to constrain the outputs of geophysical inversions to be geologically reasonable. Examples include the hierarchical Bayesian "lithologic tomography" of Bosch (1999) and Bosch et al. (2001), structure-coupled multiphysics approaches (Gallardo and Meju, 2011;de Pasquale et al., 2019), and flexible models based on Voronoi tessellations or multi-point statistics (e.g., Cordua et al., 2012;Pirot et al., 2017). Level-set methods (Osher and Sethian, 1988;Santosa, 1996;Li et al., 2017;Zheglova et al., 2018;Giraud et al., 2021a) solve for the boundaries of discrete rock units; these methods are more parsimonious than traditional volumetric inversions but still invert for a flexible, high-dimensional description of static 3-D properties.
In contrast, the core problem in geology is to interpret observations in terms of geological histories and processes (Frodeman, 1995). This becomes important in the structural geology of ore-forming systems, where details of the history become important and flexible treatment of 3-D geological structures as random fields is inadequate (Wellmann et al., 2017). The mineral exploration community has developed 3-D geological forward models that can capture much of the complexity of real systems (e.g., Perrouty et al., 2014). Implicit models (Houlding, 1994) represent interfaces between geological units as isosurfaces of a scalar field defined over 3-D space. The popular industry software package 3D GeoModeller (Calcagno et al., 2008) and the open-source code GemPy (de la Varga et al., 2019) interpolate this scalar field directly from structural geological observations by cokriging (Lajaunie et al., 1997). Kinematic models (Jessell, 1981) use a parametrized description of volume deformations produced by tectonic events; the open-source package Noddy (Jessell and Valenta, 1993) is commonly used. The recently released open-source package LoopStructural (Grose et al., 2020) combines elements of co-kriging and kinematic implicit models for additional geological richness and realism. For models with no constraints from geophysics, geological uncertainty can be quantified by generating Monte Carlo realizations (Metropolis and Ulam, 1949) of geological datasets (Pakyuz-Charrier et al., 2018a, b, 2019. In a Bayesian treatment, this amounts to drawing samples from the prior (Wellmann et al., 2017).
Conditioning geological forward models on geophysical observations makes it tractable to perform full posterior inference in an interpretable parameter space of reduced dimension. This can be done using sampling methods such as Markov chain Monte Carlo methods (MCMC; Metropolis et al., 1953;Hastings, 1970), which are becoming increasingly sophisticated and widely used in geological and geophysical modeling (Mosegaard and Tarantola, 1995;Mosegaard and Sambridge, 2002;Sambridge et al., 2012). Additionally, parametric geology naturally obeys constraints imposed by the forward geological process, and can be used for inference over that process, unlike nonparametric geophysical inversions that incorporate a static, voxelized geological model as a prior mean (Giraud et al., 2017(Giraud et al., , 2019aOgarko et al., 2021;Giraud et al., 2021b). Kinematic models are especially well suited for this approach since they represent geological hypotheses in terms of structural observations and event histories but avoid the full computational burden of dynamic process-based simulations. Performing Bayesian inference over kinematic model parameters could open the door to more general Bayesian model selection over geological histories or conceptual models.
Probabilistic inversion workflows that incorporate forward geophysics based on forward-modeled 3-D structural geology are, to date, still uncommon, owing in part to computational challenges faced by sampling algorithms. The pyNoddy code (Wellmann et al., 2016) provided a wrapper to Noddy to enable Monte Carlo sampling of 3-D block models and potential fields, but not posterior sampling of an inversion conditioned on gravity or magnetic data. The Obsidian distributed inversion code (McCalman et al., 2014) implemented a geological model of a layered sedimentary basin with explicit unit boundaries to explore geothermal potential (Beardsmore et al., 2016); it featured multiple geophysical sensors and a parallel-tempered MCMC sampler for complete exploration of multi-modal posteriors. Obsidian has been extended with new within-chain MCMC proposals inside the parallel-tempered framework (Scalzo et al., 2019) and a new sensor likelihood for field observations of surface lithostratigraphy (Olierook et al., 2020), but the limitations of its geological model make it an unlikely engine for generalpurpose inversions. Wellmann et al. (2017) present a workflow that uses GeoModeller to render the geology, Noddy for calculation of geophysical fields, and pymc2 (Patil et al., 2010) for MCMC sampling. GemPy's support for information other than structural geological measurements is as yet fairly limited, but includes forward modeling of gravity sensors and topology. This paper describes a simple Bayesian inversion framework, called Blockworlds, which to our knowledge is the first to perform full posterior inference based on forward geophysics from an implicit geological model with kinematic elements.
Although not yet intended to serve as a productionready geological modeling package, Blockworlds' design addresses an obstacle for inversion workflows that resample the parametrized 3-D geometry of geological units onto a discrete volumetric mesh for geophysical calculations. Discontinuous changes in rock properties across unit interfaces may become undersampled unless the mesh is chosen to align with those interfaces (Wellmann and Caumon, 2018); since implicit and kinematic geological models naturally represent geology volumetrically, this alignment is not easily made without first evaluating the model, incurring additional computational overhead. Flexible geophysical inversions can adjust rock properties in individual boundary voxels to partially compensate for the loss of accuracy, but models that condition rock properties directly on geological parameters lose much of this freedom. This can result in artifacts in the likelihood that hinder convergence of estimation methods, degrade posterior inference over structural parameters, and preclude the use of posterior derivative information.
To address the issue of undersampled interfaces, Blockworlds incorporates numerical anti-aliasing into its discretization step. Anti-aliasing to address undersampling has a long history in computer graphics (Crow, 1977;Catmull, 1978;Cook, 1986;Öztireli, 2020) and in geophysics, for example, finite-difference solutions of seismic wave equations (Backus, 1962;Muir et al., 1992;Moczo et al., 2002;Capdeville et al., 2010;Koene et al., 2021). Obsidian antialiases physical rock properties using volume averages in voxels with partial unit membership, though this is not mentioned in associated publications (McCalman et al., 2014;Beardsmore et al., 2016), and the implementation depends upon the Obsidian parametrization. Blockworlds' prescription is more general, operating on the scalar field values that define interface positions in all implicit models, and thus can be straightforwardly implemented for implicit models using other types of interpolation schemes such as co-kriging. This paper is organized as follows. Section 2 explains the aliasing effect, including its influence on the likelihood, and describes the algorithm used in Blockworlds to address it. Section 3 describes how Blockworlds evaluates kinematic event histories, then introduces a set of synthetic kinematic geological models we use as benchmarks. Section 4 summarizes gravity inversion experiments based on these models, demonstrating the influence of anti-aliasing on MCMC sampling. Section 5 discusses future directions, including the use of anti-aliasing with other geophysical sensors and curvilinear coordinate systems, and we conclude in Sect. 6.

Chaining forward models in geophysics calculations from structural geology
The forward modeling of observations falls under the calculation of the likelihood p(d|θ ), which describes the probability of observing data d given that the causal process generating the data has true parameters θ . A Bayesian inversion proceeds by combining the likelihood with a prior p(θ ) that describes the strength of belief, in terms of probability, that the causal process has true parameters θ before any data are taken into account. In nonparametric geophysical inversions, the prior includes regularization terms that penalize undesirable solutions. In parametric inversions, the choice of parametrization corresponds to a strong regularization on the discretized geology, with the prior distribution providing further constraints. The posterior distribution p(θ |d), describing probabilistic beliefs about θ once the data have been taken into account, is then determined through Bayes's theorem: For parameter estimation, the unnormalized right-hand side can then be sampled using MCMC algorithms for full uncertainty quantification. While some measurements, such as structural observations, can be computed directly from a continuous functional form for the geological forward model, the calculation of likelihoods based on simulation of geophysical sensors may require discretization of the geology. The likelihood terms in this case each involve the composition of two forward models: 1. a mapping g : → G from the parameter space into a space G of discrete volumetric representations (such as block models); and 2. a mapping f : G → D from the space of discretizations into the space of possible realizations of the data.
Structural geology is chiefly concerned with the map g, and often reckons uncertainty not in terms of parameter variance, but in terms of the properties of an ensemble of discrete voxelized models (e.g., Wellmann and Regenauer-Lieb, 2012;Lindsay et al., 2012). Inversions of geophysical sensor data such as gravity, magnetism, conductivity, or seismic data are chiefly concerned with the map f . The accuracy of a parametric inversion will depend on the accuracy with which f •g can be computed.
Although discretization methods such as adaptive meshes (Rawlinson et al., 2014) or basis functions aligned with geological features (Wellmann and Caumon, 2018) can improve the fidelity of g, these mappings are not guaranteed to be suitable for efficient and accurate computation of f . Furthermore, existing geological engines frequently export to a simple fixed basis that may be adequate for visualization in an 3644 R. Scalzo et al.: Anti-aliased geophysics for geological forward models interactive workflow but causes trouble in automated inversion workflows, producing the aliasing effects discussed in the next subsection.

Aliasing and its effects on the posterior
To show a straightforward example, we calculate the posterior density for the parameters of a uniform-density spherical inclusion (sphere radius and mass density) given the gravitational signal. An analytic solution exists, so we can compare directly to the true posterior, and with only two parameters we can simply evaluate over a grid of parameter values rather than using MCMC. We model a 1 km 3 cubical volume, and fix the sphere's position at the center. Gravity anomaly data were generated from the analytic model on an evenly spaced 20×20 survey grid at the surface (z = 0), with Gaussian measurement noise at the level of 10 % of the signal amplitude added. We used version 0.14.0 of the SimPEG library (Cockett et al., 2015;Heagy et al., 2020) to generate meshes and calculate the action of the forward model for the gravity sensor in the inversion loop. To emphasize the role of the likelihood in a scenario with vague prior constraints, we use uniform priors on the mass density ρ (3.0 ± 0.5) and radius R (300 ± 100 m) of the sphere. Given spherical symmetry, the data constrain only the total mass M of the intrusion, and so the region of high posterior probability density follows the curve (2) Figure 1 shows the results of this exercise. Four discretizations of the sphere are shown: a coarse mesh, with the rock density for each cell queried from the true geology at the center of that cell; a higher-resolution mesh, with cell sizes refined by a factor of 4 along each axis; a coarse mesh on which the rock density has been averaged throughout the cell, using the anti-aliasing scheme described in the next section; and a high-resolution mesh using the same anti-aliasing scheme. The gravity fields are visually indistinguishable for all four versions, but the posteriors look very different.
When inverting on the coarse mesh with no anti-aliasing, the density in a given cell changes only when the sphere radius crosses over the center of one or more mesh cells, and then it changes discontinuously (see last row, first column of Fig. 1). This in turn induces spurious structure into the likelihood, and hence the posterior distribution of the inversion parameters. The numerical posterior becomes a collection of isolated modes scattered along the locus of degeneracy (dashed red line) between ρ (y axis) and R (x axis) on which the true parameters (red cross) lie, none of which appropriately quantify the uncertainty in the true analytic problem. These artifacts can be suppressed by refining the mesh, though at the cost of increasing the number of voxels (and the computation time); in this case, by a factor of 4 3 = 64. Importantly, the underlying posterior for this high-resolution mesh is not free of discontinuities, but takes on a terraced appearance.
The inversions with anti-aliased geology result in continuous, even smooth, posteriors that trace the analytic solution and run at nearly the same speed. A closer look reveals that the coarse-mesh anti-aliased posterior is slightly offset toward the bottom left relative to the analytic curve, while the fine-mesh posterior is not. This is a product of low mesh resolution, which the anti-aliasing approximation only partially mitigates (see Sect. 2.4). Each anti-aliased interface is treated locally as a plane surface. For a given gravity signal, spheres with a higher density have a smaller radius and thus higher curvature at the interface; the departure from the true posterior will grow as the sphere radius approaches the voxel size. The bias amounts to ∼ 1 % of the sphere's mass at the true parameter values, which the data are just sufficient to detect. Thus, while anti-aliasing may incidentally improve the accuracy of geophysical field calculations, its primary benefit is to reproduce the continuous structure of the underlying likelihood.
We will later directly test the influence of aliasing on the performance of MCMC, since the capacity for full posterior inference is one of the major advantages of parametric geological inversions. However, aliasing will cause problems for a broad variety of estimation algorithms. Even for fine mesh resolution, the likelihood is piecewise constant along spatially aliased geological parameters, and so components of the likelihood gradient in those directions are zero almost everywhere in a measure-theoretic sense. Methods such as Nelder-Mead (for optimization) and Metropolis random walk (for sampling) that do not rely on posterior gradient information will seem to work with a fine enough mesh, although at greater computational cost and with no warning of the extent of aliasing effects unless run multiple times from different starting points. Gradient-based methods, such as stochastic gradient descent or Hamiltonian Monte Carlo, will fail catastrophically, since the zero gradient of an aliased likelihood will not reflect its geometry. The covariance matrix used to estimate local uncertainty in the fit from an optimization algorithm will be singular, since the components of the Hessian along aliased directions will be zero. Finally, while the gravity inverse problems we consider in this paper are linear, an attempt to invert for parametric geology based on a nonlinear sensor would face discontinuous changes in the sensitivity kernel with parameters, with unpredictable effects on parameter estimation.
One could argue that the likelihood might be less badly aliased if the data constraints were weaker. Inflating the data errors can be viewed as a form of tempering, which we would expect to merge the multiple aliased modes. Figure 2 shows the results of inflating the errors in the "coarse aliased" discretization of the sphere problem by various factors until the isolated modes in the original problem have merged. The distribution becomes fully navigable only for data errors a factor of 5-10 times larger than our original problem -that is, for Figure 1. Calculation of the posterior distribution for the radius and density of a uniform spherical inclusion from gravity inversion. Top row: cross section of 3-D rock density field for the true model from which the data are generated. Second row: simulated gravity field at the surface. Third row: residuals of simulated gravity field from the analytic solution. Bottom row: cross section through the posterior for an independent Gaussian likelihood. Discretization schemes shown (columns, left to right): coarse aliased mesh (15 3 cells), fine aliased mesh (60 3 cells), coarse anti-aliased mesh, fine anti-aliased mesh. The true parameters are shown by the red crosses, while the dashed red lines indicate the degeneracy relation in Eq. (2) that constitutes the maximum-likelihood ridge of the analytic solution.
data errors comparable to the amplitude of the signal itself. Furthermore, the blockiness or terracing is a function of the discretization scale only, so that any estimation methods relying on gradients will break down along the aliased spatial parameters in the same way as before.
This kind of catastrophic breakdown occurs because of the restrictive conditioning of rock properties on parametric geology, so we do not expect the same kind of effect to arise in flexible geophysical inversions that derive their priors from discretized geological models (e.g., Giraud et al., 2021b). The geophysical sensor model is a smooth function of the rock properties, and so, when the rock properties are the primary model parameters, the likelihood remains smooth. Highly restrictive model error terms that are inconsistent with the data may still hamper performance or inference. While it would be interesting to examine nonzero voxelization model errors for the kind of parametric inversion we investigate in this paper, this is a technically complex issue that we defer to future work. For now we note simply that aliasing still arises even in a parametric model that is in other respects perfectly specified, and that treating it explicitly will remove aliasing contributions to model errors in future parametric inversions.
While our example may seem artificial, widely used geological modeling tools such as Noddy and GeoModeller usually export on rectangular meshes, with lithology or rock properties evaluated at the center of each mesh cell precisely as described above. Any ongoing development or support of geological modeling tools intended for use in probabilistic workflows should keep examples like this in mind. Complex structure imparted by faults, folding, dykes, and sills is highly sensitive to discretization. The situation may arise where the causative body for a strong geophysical anomaly, such as a thin and relatively magnetic or remanently magnetized dyke, is removed from the geological prior due to overly coarse discretization parameters, while a resulting strong magnetic response remains. Inversion schemes are not intended to address situations where the target is unintentionally removed from the data.

Fitting an anti-aliasing function
The main idea behind anti-aliasing is to produce rock property values in each mesh cell such that the response of a sensor to the anti-aliased model best approximates the action of the same sensor on the same geology exported to a higher-resolution mesh. In this section, we will develop an anti-aliasing prescription for gravity anomaly, which responds linearly to mass density. We frame the problem as a regression that uses the position of a geological interface within a mesh cell to predict what the mean mass density would be for that cell in a high-resolution model.
The top panel of Fig. 3 shows the geometry of the problem. For simplicity, we assume constant rock density within each unit, which reduces the problem to predicting the fraction of the cell's volume lying to either side of the interface. We further assume that the interface is flat at the spatial scale of a single mesh cell, so that it can be approximated by a plane surface defined by a point r = (r x , r y , r z ) and a unit normal n = (n x , n y , n z ). Due to the translational symmetry of a plane, the influence of r can be reduced to the projected distance u ⊥ of the plane from the voxel center r 0 . The unit normal n can be described in terms of two angles in spherical coordinates; due to the rotational symmetries of a cubical voxel, it is enough to specify the polar angle with respect to the normal to the cube face that a ray drawn from r 0 along n intersects, and the azimuthal angle with respect to a coordinate system projected onto that cube face. Expressing the angles in terms of ratios of components of n results in three main predictive features: where h is the voxel size, n min = min{n x , n y , n z }, n max = max{n x , n y , n z }, and r 0 is the voxel center.
To train the regression, we generate a synthetic dataset of 1000 (r, n) pairs with which a plane interface might intersect a generic voxel. In each pair, r is distributed uniformly in space, and n uniformly in solid angle. We calculate the partial voxel volume beneath the plane numerically by mesh refinement to 20× higher resolution. Figure 3 shows three different possible functional forms for the regression: a piecewise linear function of u ⊥ alone; a linear regression model including interaction terms up to third order, where the regressors have been transformed by a hyperbolic tangent function to maintain boundary conditions and keep fractional volumes between 0 and 1; and a Gaussian process regression. A linear regression model based on the single feature u ⊥ -the normal distance to the plane from the Figure 3. (a) Numerical approach to calculating partial volumes. The interface is approximated by a plane running through the voxel. A cloud of points is generated in a regular grid, and the partial volume is then the fraction of these points that satisfy u ⊥ = (r − r 0 ) · n/ h = 0, where r 0 = (r x , r y , r z ) 0 is a point on the interface and n = (n x , n y , n z ) is a unit normal vector. (b) Regression models for fast evaluation of partial volumes using data generated from the approximate numerical scheme (gray curves): a piecewise linear interpolation over u ⊥ (red), a generalized linear model with up to third-order terms (blue), and a Gaussian process that also includes features associated with the orientation of the interface (green). (c) Precision of partial volume models. The piecewise model reaches 2 % RMS accuracy. A linear regression reaches 1 % accuracy, as does a Gaussian process regression with a single feature u ⊥ . A GP with three features that captures the directionality of the interface with respect to the cube boundaries reaches 0.3 % accuracy.
voxel center -can provide smooth anti-aliasing to 1 % RMS accuracy (5 % worst-case); the optimized form is This functional form provides negligible computational overhead in our implementation as part of a MCMC inversion loop and has derivatives of all orders. Including orientation features in the linear regression results in negligible improvements in accuracy while increasing computation time. The Gaussian process does not provide meaningful improvement over the linear regression using u ⊥ alone; it attains an accuracy of 0.3 % RMS (3 % in the worst case) using three features, but is many times slower to evaluate. We use Eq. (6) for further experiments in this paper.

Potential limitations of anti-aliasing
Since the anti-aliasing approximation amounts to a strong prior on sub-mesh structure, understanding its limitations is critical in practical modeling. The approximation giving rise to Eqs.
(3) and (6) will break down when interface curvature becomes significant. Common sense suggests that no length scale parameter in the inference problem should ever be less than the voxel size h. These situations can be difficult to detect from inside the inversion. Spatial scale parameters, such as fold wavelengths, may produce high-curvature interfaces only in some regions of parameter space. Curvature measured using the Hessian adds computational overhead and may miss structures beneath the Nyquist limit if evaluated on the same mesh as the inversion itself. However, as shown in Fig. 1, departures from the true posterior can occur even at higher resolution, depending upon the resolving power of the data and the structure of the likelihood. Another drawback in anti-aliasing is that it ignores the relative orientations of interfaces in voxels spanning multiple interfaces. If only one interface passes through a voxel, Eq. (6) should result in a faithful representation of its density. If, however, that voxel is later faulted or brought up against an unconformity, the true mean density in the voxel depends on which of the two formations is preferentially replaced by the new obtruding unit; Eq. (6) will only give some mean value that is roughly correct when the two interfaces intersect at right angles. Anti-aliasing will thus become increasingly inaccurate with each successive application to a voxel. This effect can cause biases even in geologies with exactly flat interfaces; we expect it to be more pronounced in models with multiply faulted interfaces in regions where the geophysical sensitivity is the highest, for example, near the surface. Such regions of complexity can be identified by flagging voxels within a projected distance u ⊥ < h of an interface, and the additional uncertainty in the lithology can be treated explicitly -for example, by assigning latent variables in the statistical model to account for it.

R. Scalzo et al.: Anti-aliased geophysics for geological forward models
Once an inversion has been performed, the best practice to assess bias will still be to repeat it with a different cell size. In the next section we will discuss metrics to evaluate the global agreement of posteriors based on different mesh sizes, allowing quantitative evaluation, online or offline, of the value to the user of running at a given mesh resolution.

Experiments with kinematic models
To illustrate how anti-aliasing interacts with more realistic geological structures, Blockworlds implements a simplified kinematic model inspired by Noddy (Jessell, 1981;Jessell and Valenta, 1993). We chose to write our own demonstration code rather than modifying Noddy both for ease of prototyping and to demonstrate how anti-aliasing interacts with the kinematic calculations. We construct a range of 3-D models and visualize slices through each Bayesian model posterior to demonstrate the influence of anti-aliasing on the calculations.

Kinematic model events
The action of kinematic events in Blockworlds is recursive, with each new event modifying the geology resulting from the events before it. Since the notation describing this action is complex, we include a summary of symbols defined in this subsection in Table 1.
Each event is parametrized by a collection of parameters θ i , with i = {θ 0 , θ 1 , . . ., θ i } denoting the collection of parameters for all events up to event i. We denote the function that renders 3-D rock properties based on all events up to event i by g i (r| i ). Although geology can be evaluated on a mesh of any desired resolution, in practice the recursive evaluation means that each event modifies rock properties as rendered on the same discrete mesh as the previous event.
The calculation begins with a basement layer of uniform density, g 0 (r|ρ 0 ) = ρ 0 , which has no interfaces and thus requires no anti-aliasing. The action of the anti-aliasing operator A for an interface between two sub-volumes with spatial rock property fields g + and g − at a projected distance u ⊥ from the voxel center is given by where u ⊥ and v(u ⊥ ) are given by Eqs. (3) and (6). The antialiasing must be applied in the co-moving coordinate frame of each event, so that rock properties stored on the discrete rock property mesh are recursively averaged or transported by later events (see Fig. 4). Note also that u ⊥ , and thus A, depends on the mesh spacing. The Blockworlds code covers four elementary event types, each parametrized by its own scalar field that takes a zero value at the relevant interface: -Stratigraphic layer. θ i = ( z i , ρ i ) for a layer with thickness z i and mass density ρ i , which has the ef- -Fault. θ i = (r 0,i , n i , s i ) for a fault passing through a point r 0,i with unit normal n i and dip-slip displacement s i . Our parametrization constrains r = (x, y, 0) to lie on the surface, and uses a polar representation for n derived from dip and dip direction (Pakyuz-Charrier et al., 2018b). Given these elements, we calculate v = (u z ×n)×n |(u z ×n)×n| , resulting in a unit vector pointing in the dipslip direction parallel to the fault plane, with no strikeslip component. Then g i (r| i ) =A r − r 0,i · n i ; g i−1 (r + sv| i−1 ) , which has the effect of displacing the geology in the half-space u ⊥ > 0 by the slip vector sv. -Fold. θ i = (n i , ψ i , φ i , L i , B i ) for a fold defined by a unit vector n i , pitch angle ψ i , phase angle φ i , wavelength L i and amplitude B i . Given n i , we generate two unit vectors v 0 = u z ×n |u z ×n| and v 1 = (u z ×n)×n |(u z ×n)×n| , defining an orthonormal structural frame {v 0 , v 1 , n i }. Rock elements are displaced laterally by a vector field so that the rock properties after the fold are given by -Spherical intrusion. θ i = (r 0,i , R i , ρ i ) for a sphere of radius R i and uniform density ρ i centered at r 0,i : Figure 4 shows the action of the kinematic elements of the model alongside an anti-aliased version of the voxelized density volume, calculated over a 1 km 3 rectilinear mesh 15 × 15 × 15 voxels on a side.

Setup of specific 3-D models
We construct a series of 15 kinematic models with true configurations as shown in Fig. 5. Each model occupies a 1 km 3 cubical volume, and is discretized on a rectilinear mesh with cubical voxels. All have three stratigraphic layers and two additional tectonic events, varying in order and positioning.
The first set of models focus on planar dip-slip faults, with final configurations as follows:  We also include some additional folded configurations, selected to test the anticipated limitations of anti-aliasing described in Sect. 2.4 (these use the same stratigraphic history as the two-fault events except when parameter changes are explicitly noted): We choose prior distributions for the parameters to simulate the realistic incorporation of structural geological knowledge, as shown in Tables 2 and 3. We use normal and lognormal distributions over non-negative quantities, and von Mises-Fisher distributions over angular quantities, for simplicity and interpretability in terms of scale parameters, which reflect the precision of available structural information. The positions of the anchor points lying on the fault planes have very narrow priors (σ = 1 m), appropriate for surface observation localized by GPS. The fault slip parameters reflect rough estimates (σ = 150 m), an extent comparable to the true slip value for each fault. The polar representations of the fault directions are constrained by a von Mises-    . In each case, the prior mode rests at the true parameter values to ensure that misspecification of the prior does not complicate the aliasing effects we set out to examine.
For each model, we examine four regimes of resolution and aliasing. We consider a "coarse" 15 × 15 × 15 mesh (h = 66.6 m) as well as a "fine" 75 × 75 × 75 mesh (h = 13.3 m), both with and without anti-aliasing. The anti-aliased model on the fine mesh is taken to be the true model for the purposes of calculating a common synthetic forward gravity dataset against which to evaluate the likelihood.
We generate synthetic geophysics data based on the true model parameters, with measurements spaced evenly in a 20 × 20 grid at the surface (grid spacing of 50 m), and add 3652 R. Scalzo et al.: Anti-aliased geophysics for geological forward models independent Gaussian noise with a standard deviation σ 0 of 5 % of the sample standard deviation of the data across the survey. Each data point y j is thus generated according to where f j (θ ) is the forward model for the geophysical measurement y j given the geological parameters θ . In a realistic situation, we may not know the noise variance σ 2 exactly, but we can account for our uncertain knowledge of it by including a hierarchical prior for σ 2 . If we choose this prior to be an inverse gamma distribution, the integral over σ 2 can be solved analytically, saving the computational expense of sampling over σ 2 by MCMC (see McCalman et al., 2014;Scalzo et al., 2019): so that the likelihood has the form of a t-distribution with ν = 2α degrees of freedom and scale parameter √ β/α. For our experiments, we choose α = 2.5, β = 2.5 × σ 2 0 , a weakly informative prior with a mean and mode close to the true variance used to generate the data. The full likelihood is then the product of the likelihoods for independent data points, As with the earlier synthetic examples for a spherical intrusion, we used SimPEG (version 0.14.0; Cockett et al., 2015;Heagy et al., 2020) for the calculation of forward gravity. The SimPEG Simulation3DIntegral class uses a closed-form integral expression for the contribution to the gravitational field from a rectangular prism of constant density (Okabe, 1979;Li and Oldenburg, 1998): for example, for the z-component, x i ln y i + r ij k + y j ln y j + r ij k +2z k arctan where r = (x, y, z) is the location of the gravity sensor, r ij k = (x i , y j , z k ) runs over each corner of the prism, and r = r − r ij k with r = |r |. This form is widely used in geophysics and is one of the benefits of working over rectangular meshes. The total gravity signal is then the sum of the contributions for each mesh cell. Since the expression is linear in the density for each mesh cell, the forward model sensitivities can be cached for fast likelihood evaluation.

Posterior slices and MCMC sampling
To illustrate some of the effects caused by aliasing and some of the limits of anti-aliasing, we visualize a series of twodimensional slices through the posterior distribution. Pairs of free parameters are scanned on a regular 30×30 grid centered on their true values, with all other parameters held fixed at their true values. This produces a set of figures similar to Fig. 1 for the kinematic models. We also perform MCMC over the parameters of the kinematic models to demonstrate the impact of anti-aliasing on chain mixing. The aliased posterior has no useful derivative information for structural parameters, so we are limited to random walk proposals. For our tests, we choose the adaptive Metropolis random walk algorithm of Haario et al. (2001), which is used in contemporary studies such as Wellmann et al. (2017) and is the simplest algorithm likely to give good performance on posteriors with the kind of strong covariances seen in the conditional posterior slices. This algorithm starts from an initial guess θ 0 and proposes a new value θ at time step t from a multivariate normal distribution centered on the current state θ t : where η d is a global step size parameter that only depends on the dimension of the space, and The covariance matrix of the proposal is estimated from the current chain history, so that after an initial nonadaptive random walk period of length t 0 steps, the walk gradually transitions from an initial-guess covariance 0 to match the estimated posterior covariance. States are accepted according to the usual Metropolis-Hastings criterion, setting θ t+1 = θ with probability and θ t+1 = θ t with probability 1 − α. We use an adaptation length of t 0 = 1000 and an initial diagonal covariance matrix of step sizes set to 20 % of the prior width for each parameter (0.02 g cm −3 for densities, 10 m for layer thicknesses, 30 m for fault slips, and 3 • for angular variables).
We use the coarse 15×15×15 mesh for calculations since this will give much faster results for MCMC and highlight the impact of aliasing on chain mixing. We run M = 4 chains for N = 10 5 samples from the posterior of each model, with and without anti-aliasing. We initialize each chain with an independent prior draw and discard the first 20 % of samples as burn-in. Each chain took 6 min to run on an Intel Core i7 2.6 GHz processor.
We use the integrated autocorrelation time to measure MCMC efficiency within each chain, and the potential scale reduction factor (PSRF; Gelman and Rubin, 1992) to estimate the extent of convergence to the true posterior. The autocorrelation time τ is the approximate number of chain iterates required to produce a single independent sample from the posterior, so that the chain length N divided by τ becomes one possible measure of effective sample size. The potential scale reduction factorR, formed from multiple chains, is the factor by which the posterior variance estimated from those chains could be further reduced by continuing to sample. A PSRF near 1 suggests that multiple chains from different starting conditions have achieved similar means and variances and thus are sampling from a common stationary distribution; a large PSRF shows a residual dependence on starting conditions. The PSRF is usually given for each parameter in a multi-dimensional chain; some dimensions may mix quickly, while others take a long time to converge.
Finally, to quantify the overall accuracy of the posterior density calculation, we use the Kullback-Liebler divergence where p(θ) = p lo (θ |y) represents the posterior based on a low-resolution mesh (with or without anti-aliasing), and q(θ ) = p hi (θ |y) represents the high-resolution, anti-aliased posterior. D KL is a global measure sensitive to biases and covariances across all parameters. Since it requires integration over the posterior, this metric is not available to check results of single-point optimization algorithms. We use the MCMC chain itself to calculate the integral, representing it as the difference in log probability density averaged over samples from the low-resolution posterior.

Posterior cross sections
Two-dimensional slices through the posterior distribution for selected variable pairs are shown for Model 1 (in Fig. 6) and Model 6 (in Fig. 7). Although we do not include similar figures for every model in the main text, equivalent sets of plots for all models can be regenerated automatically by scripts in the Blockworlds repository. Model 1 is relatively well behaved; the aliased posterior has a single dominant mode near the true value. This is a situation in which the aliasing effects function mainly to make the likelihood appear blocky and terraced. Although proposals that require derivatives of the likelihood will fail on this posterior, it can still be navigated by appropriately tuned random walks, or by the discontinuous Hamiltonian Monte Carlo sampler of Nishimura et al. (2020).
Nevertheless, the benefits of anti-aliasing are still clear. Merely increasing the mesh resolution of the coarse aliased model by a factor of 5 (and the computational cost by over 100×) is insufficient to completely suppress aliasing artifacts. The coarse aliased model fails to capture the full extent of the very strong correlation between the dip slips of the two faults, though the fine aliased model does so reasonably well. The posteriors of the two anti-aliased models deviate only negligibly from each other for all parameters.
Model 6 presents a more challenging case where a coarse mesh does not fully resolve structures close to the surface, resulting in bias for the low-resolution models. The layer thicknesses of the coarse aliased model are explicitly multimodal as the interfaces hover between depths, and the modes for several variables are significantly offset from their true values. These values become sharp modes in a broader parameter space, easily missed by any inappropriately scaled MCMC proposal. The fine aliased model reproduces the overall posterior shapes, with some distortions relative to the fine anti-aliased reference model. The coarse anti-aliased model produces a smooth posterior and recovers the correct overall correlations between parameters, but are somewhat offset from the true values.
The differences between low-and high-resolution models are, predictably, more dramatic for folded models where the priors do not exclude fold wavelengths close to the Nyquist limit for the coarse mesh scale, such as Models 10 and 11. Other models, such as Model 9, show that anti-aliasing gives much better results for fold wavelengths greater than the mesh scale. In Model 11, where the undersampled structure is positioned farther beneath the surface, the apparent effects are less apparent than in Model 10, but still appear in directly linked variables such as the fold wavelength.
The biases caused by aliasing seem to be strongest in angular variables such as fault directions. For Model 6, the bestfit dip-slip angles are offset from the true values by more than 10 • , corresponding to the extent through which the scan must sweep to obtain any difference in rock property for a few near-surface voxels in the coarse aliased model. Antialiasing the coarse-mesh model largely, but not completely, mitigates this bias. In Model 10, the layer thicknesses are the only model parameters that can be robustly estimated without refining the mesh to resolve structure.
These results confirm that our algorithm can reproduce the continuous behavior in an underlying posterior distribution, and can enable significant computational savings even for complex models as long as the interfaces are smooth on the mesh scale. It cannot -and is not intended to -recover more complex structures beneath the mesh scale that violate the core assumptions under which it was derived. Figure 6. The log posterior distribution of pairs of parameters, denoted by θ , in kinematic Model 1, relative to the maximal value log 10 [p(θ |d)/p(θ MAP |d), whereθ MAP = argmax θ p(θ |d). In each slice, the true values of gridded parameters are marked with red crosses, and parameters that are not plotted are fixed at their true values. Columns correspond to different discretization schemes; from left to right: coarse mesh (15 3 ) with no anti-aliasing; fine mesh (75 3 ) with no anti-aliasing; coarse mesh with anti-aliasing; fine mesh with anti-aliasing. The top row shows a vertical cross section at y = 0 of the 3-D rock density field under each discretization scheme. Rows 2-5 correspond to different pairs of parameters in θ ; for row 2, θ consists of the two non-basement layer thicknesses; for row 3, θ consists of the dip direction angles of the two faults; for row 4, θ consists of the fault slips; for row 5, θ consists of the dip of the first fault and the thickness of the top layer.  Table 4 shows the average, best-case, and worst-case values of τ andR across all model parameters, as well as the K-L divergence D KL . The posterior is challenging to navigate and all chains have autocorrelation times of the order of hundreds to thousands of samples, comparable to the performance of Obsidian for the Moomba models (Scalzo et al., 2019). Chains for anti-aliased models have autocorrelation times that are shorter by a factor of 2-10 than chains for aliased Table 4. Performance metrics for each MCMC run: mean (best-case, worst-case) integrated autocorrelation time τ in samples (×10 3 ); mean (best-case, worst-case) potential scale reduction factorR calculated across four runs starting from independent prior draws; and sample Kullback-Liebler divergence D KL between the given model and the fine-mesh anti-aliased version, calculated by combining samples from all four runs thinned by a factor of 1000. Runs marked with "AA" are coarse-mesh anti-aliased, while those marked with only a single-digit model number have anti-aliasing turned off. models, owing to the smoother, more Gaussian structure of the posterior. Most dramatically, the aliased chains still have large PSRF for several variables, even after 10 5 samples. The anti-aliased chains converge more consistently to the same distribution, with final PSRF values close to 1 for all variables for most models, and much closer to 1 than the aliased chains even for the more challenging, adversarial model configurations. Models 5, 6, 10, 14, and 15 are challenging models for which the anti-aliased chain also has trouble mixing fully. An example trace plot from Model 5 is shown in Fig. 8. One of the chains in the anti-aliased version has fallen into a dif- ferent mode from the other three, characterized by shifts in the thicknesses and densities of stratigraphic layers as well as a shift in the dip angle of the fault. In contrast, none of the four traces in the aliased model seems to be sampling from the same distribution; each chain has learned a different proposal scaling, and chains occasionally jump between modes at adjacent parameter values characteristic of the ones seen in lower-dimensional projections. Each chain, considered on its own, would give a very different impression of the uncertainty in the inversion. This qualitative behavior is characteristic of sampling for the other models.

MCMC sampling
Complementing the trace plots and MCMC performance metrics as a global measure of posterior inference accuracy, the K-L divergence shows that anti-aliasing usually improves correspondence with the reference posterior (the highresolution, anti-aliased posterior) by a large margin. Antialiased models with large values of D KL correspond well with sub-Nyquist structures or with regions of complexity close to the surface, confirming our intuition that inference in these models benefits from a finer mesh resolution. The aliased posteriors show differences from the high-resolution posteriors spanning many orders of magnitude in probability density, averaged over those modes the chain was able to reach. Since the aliased chains did not in general mix well, their values of D KL should be taken as indicative, whereas the anti-aliased chains with lowR represent more robust estimates of the true divergence.
To fully characterize uncertainty in truly multi-modal problems, parallel tempering should be used to sample from all modes in proportion to their posterior probability. However, without anti-aliasing, the low-temperature chain in the tempering scheme will mix poorly and will rely more heavily on swap proposals to explore the posterior. The modes that do appear in anti-aliased posteriors are also more likely to represent distinct interpretations, rather than poorly resolved strong correlations between variables.

Discussion and future work
We have experimentally demonstrated the use and benefit of anti-aliasing only for kinematic models acting on a regular rectangular mesh. However, we see several natural extensions to this work that extend the range and impact of antialiasing for parametric geological models. We describe these future directions in the following section.

Anti-aliasing in terms of the implicit scalar field
Although Blockworlds forward models geology through the action of parametrized tectonic events, it is still an implicit model in the sense that the unit interfaces are defined to be level sets of a scalar field, which we parametrize simply in terms of the distance to the interface. The mathematics of our anti-aliasing prescription generalizes to any differentiable scalar field, including those used in co-kriging models that interpolate structural measurements (Lajaunie et al., 1997;Calcagno et al., 2008). These models will have different posterior geometries than kinematic models, since the modification of structural parameters produces more localized changes in the block models. However, we see no intrinsic reason why the following treatment for general scalar fields should not work just as well as for other types of implicit models, provided the interpolated geological structures have minimal curvature on the mesh scale.
For implicit geological models, each interface is defined as the level set (r) = 0 of some scalar field . The value of corresponds roughly to depth or to geological time, but has no intrinsic physical meaning except to ensure that the geologic series of interfaces it defines are conformable. In interpolation models, the specific value 0 representing an interface is fixed by observations (from surface measurements or boreholes). In this case, u ⊥ can be calculated easily by locally calibrating the scalar field to represent a physical distance in the neighborhood of an interface: for a voxel centered at r 0 , a first-order approximation of gives which, using Eq. (3) and writing the interface unit normal as n = ∇ /|∇ |, can be rearranged to give In this way, each voxel can be anti-aliased simply by evaluating the scalar field and its gradient at the voxel center. For co-kriging models, the scalar field itself must be evaluated at the voxel centers to render it, and the field gradient can be evaluated straightforwardly using the same trained model by differentiating the kernel.

Anti-aliasing in curvilinear coordinates
The development of the anti-aliasing mapping in Sect. 2.3 and its generalization in Sect. 5.1 assume that each voxel is a cube. However, mesh cells with varying aspect ratios are common in geophysical inversions, for example in dealing with varying sensitivity with depth or with boundary conditions where high resolution is less important. It may also be useful for some problems to work in curvilinear coordinates. Another useful avenue of future work would be to extend Eq. (23) to more general coordinate systems. Let J = ∂r/∂u be the Jacobian of the transformation from dimensionless coordinates u on a unit voxel to physical coordinates r, so that to first order u = J −1 (r −r 0 ). Equation (3), defining the projection of a reference point in the voxel onto an interface running through the voxel, then becomes and Eq. (3), defining u ⊥ in terms of scalar field values for an implicit model, becomes

Anti-aliasing for other geophysical sensors
In this paper we have focused on gravity, since it is among the most widely used geophysical sensors in an exploration context and its linear response makes an anti-aliasing treatment straightforward. All of our results extend immediately to magnetic sensors given the strong mathematical equivalence, and may also be appropriate for linear forward models of other sensors such as thermal or electric conductivity, or for the slowness in travel-time tomography. Useful schemes to anti-alias forward-modeled geology may exist for other sensors, as long as the sensor action at scales smaller than the mesh spacing can be usefully approximated by a computationally simple function of the rock properties and the interface geometry. Frequency-dependent sensors that probe a range of physical scales may require a frequency-dependent anti-aliasing function, and similarly for sensors with anisotropic interactions with interfaces. While there is nothing fundamental that prevents anti-aliasing for sensors that respond nonlinearly to rock properties, mesh refinement may be more important in these cases to ensure the numerical accuracy, as well as the continuity, of the posterior. The framework in this paper treats only quantities defined at mesh cell centers; while it may be possible to derive similar relations for quantities defined on mesh cell faces and edges, as in finite-volume treatments of electromagnetic sensors, this represents future work.

Conclusions
Our experiments show the potential pitfalls of using oversimplified projections of 3-D structural geology onto a volumetric basis for calculation of synthetic geophysics, and demonstrate an intuitive, efficient solution. Anti-aliasing reproduces the smooth behavior of the underlying posterior with respect to the geological parameters to enhance the convergence of optimization or the mixing of sampling methods and to enable the use of derivative information in these methods. Antialiasing enables a fixed volumetric basis to more faithfully represent sensor response to latent discrete geology with interfaces that are flat at the scale of a mesh cell, reducing the computational burden of forward geophysics in an inversion loop. Our algorithm is also coordinate invariant and can be combined with curvilinear meshes, and with mesh refinement techniques such as octrees (e.g., Haber and Heldmann, 2007), for even stronger results. Finally, the K-L divergence between versions of the log posterior using different mesh resolutions can be calculated using MCMC samples either online or in post-processing as a diagnostic, allowing the user to determine how much the problem would benefit from additional mesh refinement.
We have focused on calculations over 3-D Cartesian volumetric meshes because implicit and kinematic models are naturally volumetric, and because existing geological modeling codes already export rock properties onto meshes as block models. Introducing anti-aliasing into these geological models is a minimally invasive modification to enhance their use in MCMC-based Bayesian inversions. Other discretization schemes built to align with geological interfaces, such as pillar grids (Wellmann and Caumon, 2018), may mitigate aliasing along the pillar, but still suffer aliasing of features cutting across the pillar axis and are more limited in their representational power. Alternate methods for calculating gravity and magnetics, such as the Cauchy surface method (Zhdanov and Liu, 2013;Cai and Zhdanov, 2015) or finitevolume methods on tetrahedral meshes (Götze and Lahmeyer, 1988;Schmidt et al., 2020), require many fewer discrete elements than a 3-D volumetric mesh to achieve a given accuracy, since they follow an explicit lower-dimensional discretization of each interface. These methods are most suitable for geological models that already parametrize unit in-terfaces explicitly. Using them with implicit models, or with kinematic models that work by deforming and transporting volumetric elements, would introduce additional computational overhead in tracking the positions of interfaces and updating an associated lower-dimensional mesh. The sensor response coefficients would then also have to be updated each time the model parameters are updated. There may still be geological use cases where tracking interfaces during the sampling process is more efficient than a finite volume solution involving every voxel, but we defer the examination of such cases to future work.
The accurate and efficient projection of these simple geological models onto meshes for geophysics calculations is a prerequisite to inversion for structural and kinematic parameters in more realistic situations. We can now take clear next steps towards MCMC sampling of kinematic histories for richer, higher-dimensional models. In addition, we can now move towards the sampling of hierarchical geophysical inversions that use a parametric structural model as a mean function. This will enable voxel-space inversions for which the geological prior is expressed in terms of uncertain interpretable parameters, or inversions for geology that include uncertainty due to residual rock property variations constrained by geophysics. These would represent more complete probabilistic treatments of geological uncertainty in light of available constraints from geophysics.

Appendix A: MCMC performance metrics
In the definitions to follow, we follow notation introduced in Scalzo et al. (2019). Consider M Markov chains, each of length N and with d-dimensional parameter vectors, which are run independently or in parallel. Let θ [j ] ki denote parameter k of d, drawn at iteration [j ] of N in chain i of M. Let denote the sample mean of parameter k over the N iterates of chain i, and denote the sample variance of parameter k over chain i. Let ki denote the sample mean ofθ ki across all chains. To summarize the variation in parameter estimates across chains, let denote the sample variance inθ ki across chains, and denote the sample mean of s 2 ki across chains. Thus, B k is a measure of the variance in parameter estimates made from the history of samples in any single chain on its own, and W k is a measure of the overall variance of a parameter within any single chain.

A1 Integrated autocorrelation time
The autocorrelation function, measuring the correlation between parameter draws separated by a lag l when treating each chain as a time series, is The integrated autocorrelation time (IACT), the sum of the autocorrelation function over lags l, gives an estimate of the number of chain samples required to obtain an independent draw from the stationary distribution: A2 Potential scale reduction factor The potential scale reduction factor (PSRF) was introduced as a metric for chain convergence by Gelman and Rubin (1992). It measures the extent to which uncertainty in a parameter could be reduced by continuing to sample beyond the nominal chain length N. A simple version that is often implemented iŝ The behavior of the metric is driven by the ratio B k /W k , or the variance in parameter means from different chains as a fraction of the overall marginal variance in that parameter. This can be large because of small number of statistics in a unimodal problem, but it can also be large because of poor mixing between modes in a multimodal problem, which increases the autocorrelation between samples. Gelman and Rubin (1992) advocate searching for multiple modes in advance and drawing starting points for MCMC from an overdispersed approximation to the posterior. They also account for sampling variability, including correlation between samples, via the modified metriĉ where ν = 2V 2 /v ar(V ) represents the number of degrees of freedom in a t-distribution forθ k , witĥ × cov s 2 ki ,θ 2 ki − 2θ k cov s 2 ki ,θ ki .
We use the metricR for our experiments, given that autocorrelation times can be quite long for our chains.
Code and data availability. The version of the Blockworlds model code used in this paper is archived with Zenodo (https://doi.org/10.5281/zenodo.5759225; Scalzo, 2021). The datasets used in the inversions are synthetic and can be reproduced exactly by running a set of scripts from the command line that fix the random number seed used to generate those datasets, as described in the package manual.
Author contributions. The study was conceptualized by SC, who, with MJ, provided funding and resources. RS was responsible for project administration, designed the methodology, wrote the Blockworlds code resulting in v0.1, and carried out the main investigation and formal analysis under supervision from SC, EC, and MJ. RS and ML validated and visualized the results. RS wrote the original draft text, with contributions from ML, for which all coauthors provided a review and critical evaluation.
Competing interests. The contact author has declared that neither they nor their co-authors have any competing interests.
Disclaimer. Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Financial support. This research has been supported by the Australian Research Council (grant nos. IC190100031 and DE190100431).
Review statement. This paper was edited by Lutz Gross and reviewed by Florian Wellmann and two anonymous referees.