the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Minimumvariancebased outlier detection method using forwardsearch model error in geodetic networks
Utkan M. Durdağ
Geodetic observations are crucial for monitoring landslides, crustal movements, and volcanic activity. They are often integrated with data from interdisciplinary studies, including paleoseismological, geological, and interferometric synthetic aperture radar observations, to analyze earthquake hazards. However, outliers in geodetic observations can significantly impact the accuracy of estimation results if not reliably identified. Therefore, assessing the outlier detection model's reliability is imperative to ensure accurate interpretations. Conventional and robust methods are based on the additive bias model, which may cause typeI and typeII errors. However, outliers can be regarded as additional unknown parameters in the Gauss–Markov model. It is based on modeling the outliers as unknown parameters, considering as many combinations as possible of outliers selected from the observation set. In addition, this method is expected to be more effective than conventional methods as it is based on the principle of minimal variance and eliminates the interdependence of decisions made in iterations. The primary purpose of this study is to seek an efficient outlier detection model in the geodetic networks. The efficiency of the proposed model was measured and compared with the robust and conventional methods by the mean success rate (MSR) indicator of different types and magnitudes of outliers. Thereby, this model enhances the MSR by almost 40 %–45 % compared to the Baarda and Danish (with the variance unknown case) method for multiple outliers. Besides, the proposed model is 20 %–30 % more successful than the others in the lowcontrollability observations of the leveling network.
 Article
(1046 KB)  Fulltext XML
 BibTeX
 EndNote
Conventional tests for outliers and robust M estimation are based on the least squares estimation (LSE). If an observation contains an outlier, the LSE method ceases to be the optimal estimation method in terms of a minimumvariance unbiased estimator. Once outliers are detected and isolated, the LSE can be called an efficient estimation. Otherwise, an undetected outlier has a slight deviation from the normality assumption that may cause a smearing effect on all estimated parameters regardless of whether using LSE directly or indirectly which may be named the local influence function (IF) of the LSE (Gao et al., 1992; Hekimoglu et al., 2010; Nowel, 2020). For different bias intervals, the smearing effect of the LSE that behaves systematically as a function of partial redundancy has been proven by Durdag et al. (2022). Normalized residuals, which would be exposed to the smearing effect, are investigated to identify and isolate outliers by conventional tests for outliers and some robust methods such as M estimation (Zienkiewicz and Dąbrowski, 2023; Wang et al., 2021; Batilović et al., 2021). Thereby the falsified test result may induce typeI errors. In addition to the low efficiency of LSE results, the low success of the F test shown by Hekimoglu led researchers to seek a more reliable and effective method such as the univariate method and original observations (Hekimoglu, 1999; Erdogan, 2014; Hekimoglu et al., 2014). Although these methods boost the reliabilities of the conventional methods, the identification of outliers in these models is based on the same procedure as the conventional and robust methods. If the normalized residual exceeds 3 times its standard deviation (SD), also called the 3σ rule, an observation is flagged as an outlier (Lehmann, 2013). However, tests for outliers can be dealt with sufficiently with a single outlier since the LSE has an unbounded IF (Duchnowski, 2011; Maronna et al., 2019; Huber, 1981; Durdag et al., 2022). Studies show that the reliability of these techniques, established with the additive bias model, decreases significantly as the number of outliers increases. In the decision stage, the outliers that mask or swamp other observations can produce a typeI error (false negative) and typeII error (false positive). Multiple outliers can be identified at most as the number of possible outliers (${m}_{\mathrm{max}}\le \frac{nu}{\mathrm{2}}$) by repetitive test procedures (Hekimoglu, 2005). However, the efficiency of conventional tests is rather small when the outlier value is close to the critical value, i.e., when small outliers lie between 3–6σ.
If the rate of successful detection of an outlier using conventional and robust methods is 50 % and one outlier is determined incorrectly, the probability of correctly determining two outliers remains below 50 %. This condition is based on the interdependence of each iteration. Incorrect determination at each step also reduces the possibility of identifying more than one outlier in the next step. Therefore, besides modeling the outliers as unknown, the proposed method is based on two essential factors: the principle of the slightest variance and the assumption of looking at all points with suspicion in each iteration. It has been proven by Hekimoglu et al. (2015) for linear regression that the method with which the outlier is modeled as an additional unknown gives more successful results than the robust methods. The method suggests carrying outlier detection until all possible combinations are investigated. In the ${C}_{k}^{n}$ combination, observation(s) is (are) included as an additional unknown parameter(s) in the proposed model. Then, observations are viewed with suspicion considering combinations of k elements (groups of two, three, and so forth) selected from a set of n elements (${C}_{k}^{n}$), where n is the number of observations and k denotes the number of outliers. The observation with the smallest variance among ${C}_{\mathrm{1}}^{n}$ combinations is determined. Considering the ${C}_{\mathrm{2}}^{n}$ combinations, the pair of observations was regarded as a model error and the two observations with the smallest variance were flagged as candidates. All possible combinations will be regarded until the maximum number of burdened observations that would occur up to onehalf of the degrees of freedom ($\cong f/\mathrm{2}$) for the geodetic network. The potential observations are clustered separately and compared with the specified critical values for each combination step. The model errors of the potential outlier(s) exceeding the critical value are flagged as suspicious for each combination step. The test values of all potential outliers must exceed the critical value for each combination step, and, if not, the previous candidates are detected as outliers.
The primary purpose of this study is to apply the proposed outlier detection method to geodetic networks and to establish its efficiency. The suggested model was compared with the robust methods by the mean success rate (MSR) indicator of different types and magnitudes of outliers. As in the classic models, the number of outliers is inversely proportional to the success of the presented method. When outliers have various magnitudes (e.g., small, large, gross, and extreme outliers) and specific observations are not available in the network (observations with low controllability), it has been found that the proposed method is quite successful compared to the conventional and robust methods.
Let A_{nxu} be a design matrix which has full column rank, i.e., rank (A)=u; P a weight matrix of the observations; x_{ux1} a vector of the unknown parameter; l_{nx1} an observation vector; ${{\mathbf{C}}_{ll}}_{nxn}$ an a priori covariance matrix of observations; ${{\mathbf{Q}}_{ll}}_{nxn}$ a weighted coefficient matrix of observations; and ${\mathit{\sigma}}_{\mathrm{0}}^{\mathrm{2}}$ an a priori variance factor, where n and u are the number of observation and number of unknowns, respectively. By adding v_{nx1} as a residual vector, one can get $\widehat{\mathit{x}}$, an estimated vector of unknown parameters presented in the following Gauss–Markov model (Koch, 1999):
where Q_{xx} denotes a cofactor matrix of the unknown parameters and Q_{vv} implies the cofactor matrix of the residuals.
2.1 Test for outliers
In Geodesy, procedures for the outlier detection were developed by Baarda (1968) and Pope (1976). If the observations come from the normal distribution, they are called good observations, whereas the burdened observations that contain outliers originate from another distribution. Let l_{i} be a burdened observation that has δl_{i} as an outlier; the null hypothesis,
is tested. If the observations are uncorrelated and the variance ${\mathit{\sigma}}_{\mathrm{0}}^{\mathrm{2}}$ is known, the normalized residuals can be written as
where w_{i} is the test value and q_{vv} is the cofactor of the residual for i=1…n. This is known as Baarda's method (i.e., a datasnooping test). A posteriori variance (${m}_{\mathrm{0}}^{\mathrm{2}}$) is calculated in Pope's method given by
where τ_{i} is the test value. The observation with the biggest normalized or studentized residual is tested in one loop of the iterations. The test for outliers is used iteratively if the observations contain more than one outlier. The flagged observation is removed when H_{0} is rejected. The remaining observations are adjusted once more. Until no more outliers are found, this process is repeated. However the multiple outliers cause swamping or masking effects that make it impossible to distinguish the burdened observations from the good ones. In the following sections, the robust and the proposed methods will be demonstrated to prevent the smearing effect of the LSE.
2.2 Robust methods
M estimation (Huber, 1964) is a generalized form of maximum likelihood estimation. In this paper M estimation from the Huber and Danish methods, commonly chosen to handle outliers in robust statistics, was used to compare the results of the proposed method.
2.2.1 M estimation
The reweighted LSE is applied iteratively to the nonlinear normal equation of the M estimation as follows:
where ${\widehat{\mathit{x}}}^{r}$ equals the $\widehat{\mathit{x}}$ from Eq. (2) for the first iteration and E stands for a unit matrix. r implies a number of iterations and is chosen as 5 in this paper. The weight function of Huber's M estimation is given as follows:
and the weight function of the Danish method is given by
where ${\stackrel{\mathrm{\u203e}}{v}}_{i}$ is the residual and c is taken as 1.5σ_{0}. After the diagonal elements of the $\stackrel{\mathrm{\u203e}}{\mathbf{W}}$ weight matrix are determined, ${\stackrel{\mathrm{\u203e}}{\mathit{v}}}^{r}$ and ${\widehat{\mathit{x}}}^{r}$ are recalculated for each iteration. The residual that is computed in the final iteration is detected as an outlier if it exceeds 3σ_{0}.
The Gauss–Markov model (Eq. 1) is now expanded by the u×1 vector ϵ of additional unknown parameters, also with the n×u design matrix M:
where the variance σ^{2} stands for the unit weight of the augmented model and the vector ϵ contains the outliers which are subtracted from the observations. If only the outlier Δ_{j} is present in the observation l_{j}, then one should define ϵ=Δ_{j} and M=e_{j}, where ${\mathit{e}}_{j}=\left[\mathrm{0},\mathrm{\dots},\mathrm{0},\mathrm{1},\mathrm{0},\mathrm{\dots},\mathrm{0}\right]$ for j=1…n. The jth component of e_{j} gets the value 1. For the jth observation with $\mathbf{A}=[{\mathbf{A}}_{\mathrm{1}},\mathrm{\dots},\phantom{\rule{0.125em}{0ex}}{\mathbf{A}}_{j},\mathrm{\dots}{]}^{\mathrm{T}}$, the observation equation is given as
where ${\mathbf{A}}_{j}^{\mathrm{T}}$ is the jth row vector of A and for the remainder of the observations ${\mathit{l}}_{k}+{\mathit{v}}_{k}={\mathbf{A}}_{k}^{\mathrm{T}}\widehat{\mathit{x}}(k=\mathrm{1},\mathrm{2},\mathrm{\dots},n)$, k≠j. If the outliers exist in the observations, ϵ and M are rewritten as follows:
The estimate of unknown parameters of the augmented model can, therefore, be expressed as follows (Koch, 1999):
where
The residuals are expressed for the Gauss–Markov model in Eq. (1) by
whose righthand side can be replaced in Eq. (21) as follows:
and considering Eq. (20) the following equation yields
3.1 Testing procedure
The alternative hypothesis, in the case of the presence of outliers, takes the form against the null hypothesis as follows:
One should consider all possible combinations of potentially burdened observations for the correct specification of the alternative hypothesis (Teunissen, 2006). All potential alternative hypotheses ${C}_{b}^{n}$, where n is the number of observations and b is the number of potential outliers, are considered in the detection step. Firstly, the observations are assumed to be unknown one by one in the model. The additional unknowns of the model $\widehat{\mathit{\u03f5}}$ are calculated by rewriting the relevant rows for each observation in the coefficient matrix iteratively. The design matrix can be rewritten as follows by including a dimension in the model as an unknown:
where A^{b,i} denotes the matrix of coefficients for $b=\mathrm{1},\mathrm{\dots},f/\mathrm{2}$ and $i=\mathrm{1},\mathrm{\dots},n$.
3.1.1 Calculation steps for model error
The rows of the additional column vector are rewritten iteratively for each observation, and the corresponding one is modeled as an unknown using calculation steps given below.

After calculating the cofactor matrix, the unknowns matrix is obtained:
$$\begin{array}{}\text{(27)}& {\displaystyle}{\mathbf{Q}}_{xx}^{b}=({{\mathbf{A}}^{b}}^{\mathrm{T}}{\mathbf{PA}}^{b}{)}^{+},\text{(28)}& {\displaystyle}{\widehat{\mathit{x}}}^{b}={\left({{\mathbf{A}}^{b}}^{\mathrm{T}}{\mathbf{PA}}^{b}\right)}^{+}{{\mathbf{A}}^{b}}^{\mathrm{T}}\mathbf{P}\mathit{l}.\end{array}$$ 
To determine the observation that gives the smallestvariance value, the step of calculating the residuals is given by
$$\begin{array}{}\text{(29)}& {\mathit{v}}^{b}={\mathbf{A}}^{b}{\widehat{\mathit{x}}}^{b}\mathit{l}.\end{array}$$ 
The posteriori variance is calculated as
$$\begin{array}{}\text{(30)}& ({\mathrm{s}}^{b}{)}^{\mathrm{2}}={{\mathit{v}}^{b}}^{\mathrm{T}}\mathbf{P}{\mathit{v}}^{b}/{\mathit{f}}^{b}.\end{array}$$ 
Determining the observation with minimum variance,
$$\begin{array}{}\text{(31)}& j=min({s}^{b}{)}^{\mathrm{2}}.\end{array}$$ 
After the relevant observation is determined, the test value is calculated as given by
$$\begin{array}{}\text{(32)}& T={\widehat{\mathrm{\Delta}}}_{j}/\left({s}_{\mathrm{0}}\sqrt{{\mathit{q}}_{jj}}\right).\end{array}$$
Thus, the unitweighted posteriori variances for each additional unknown parameter are calculated by
where ${u}_{k}=u+\mathrm{1}$ represents the number of the unknowns calculated for the model given in Eq. (15). The number of elements in the set of the posterior variances calculated for each observation appears as ${C}_{\mathrm{1}}^{n}$. After the acceptance or rejection of the H_{0} hypothesis is evaluated in the identification phase mentioned below, the decision is made to rewrite the model, where the unknowns are expanded for the observations two by two for the ${C}_{\mathrm{2}}^{n}$ combination. It is identified to which observation the smallestvariance value $min\phantom{\rule{0.125em}{0ex}}\mathit{\left\{}{\widehat{s}}_{i}^{\mathrm{2}}\mathit{\right\}}$ belongs, and the unknown of the relevant observation is compared with the critical value. When $min\phantom{\rule{0.125em}{0ex}}\left\{{\widehat{s}}_{i}^{\mathrm{2}}\right\}={\widehat{s}}_{k}^{\mathrm{2}}$, the absolute value of T is compared with the t test. If $\leftT\right\ge {t}_{f\mathrm{1},\mathrm{1}\mathit{\alpha}}$, H_{0} is rejected and the kth observation is flagged as an outlier. If the null hypothesis is accepted, the process ends. The model is expanded for another alternative hypothesis which assume two potential blunders in case H_{0} is rejected. The coefficient matrix is rewritten for each combination of ${C}_{\mathrm{2}}^{n}$ as
An important point to be emphasized here is that all combinations are taken into account independently of the previous result (i.e., regardless of the biased observation flagged in the previous step). For example, all potential ${C}_{\mathrm{2}}^{n}$ combinations are considered, neglecting the previous result where the kth observation was flagged. The test values of model errors (Δ_{i}, Δ_{j}) for i=1…n and j=1…n and i≠j, which have the smallest variance, are compared with the ${t}_{f\mathrm{1},\mathrm{1}\mathit{\alpha}}$ threshold value where α=0.05, whether the model errors of the observations that give the smallestvariance value are higher than the critical value or not. If both are greater than the critical value, the relevant observations are flagged as outliers. Possible ${C}_{\mathrm{3}}^{n}$ combinations are sought, and the coefficient matrix is rewritten as follows:
If all three values of unknowns exceed the critical value, they are flagged as outliers. This process is repeated for four or more combinations until all the combinations of potentially burdened observations have been considered. The $\widehat{\mathit{\u03f5}}$ vector of the observations corresponding to the minimumvariance value calculated for each combination step is compared with the critical value. If at least one of the relevant unknowns of the observations does not exceed the critical value, the H_{0} hypothesis is accepted and the observations flagged in the previous step (i.e., the latest rejected H_{0}) are approved as outliers. The flowchart of the FSME (forward search of model error) model is presented in Fig. 1.
Two cases which the variance is known and unknown were considered for robust methods as follows: ^{*} The a priori variance is known. Here, the c was taken to be 1.5, where c=1.5σ_{0} and σ_{0}=1. When the residual from the robust techniques exceeded the 3σ_{0} threshold value, it was regarded as an outlier. In the case where the a priori variance is known, it can be seen in Table 1 that the MSRs of the robust methods are higher by 0.001 than the Baarda test with α considered. Pope's test had a lower MSR than Baarda's test. However, the MSRs of the FSME are higher than the robust methods in both cases where the a priori variance is known and unknown. ${}^{**}$ The a priori variance is unknown. The standard deviation from the first iteration (LSE) was obtained for robust methods. So the c was taken from 1.5m_{0}. α was chosen as 0.05 for the Pope's test, which had a lower MSR than Baarda's. Except for the Danish* method, all other models identified an excellent observation as an outlier with a risk ranging from 0.01 % to 5 % if there was no outlier in the observations. The a posteriori variance negatively impacted the robust method's results, and the outlier's spoilt variance significantly contributed to the false detection. The a priori variance significantly affects how reliable the procedures are.
In statistics, there are different indicators to measure the reliability of tests and estimators. Hekimoglu and Koch (2000) showed that a finitesample breakdown point determined the global reliability of an estimator and a test procedure. Using the power function of the global test, a capacity in deformation networks is explored as suggested by Niemeier (1985). Also, it has been shown that the MSR results of the two testing procedures (χ^{2} and F test) are identical to their respective test powers known beforehand (Aydin, 2012). MSR depends on the number of outliers, the magnitude of an outlier, the number of unknowns, the number of observations, and the type of outliers. Since it considers these different cases, MSR is more reliable, whereas the power of the test is the same for all disparate conditions. Also, Erdogan et al. (2019) has proven that the MSR is the empirical estimation of the power of the test in outlier detection. In this study, therefore, MSR is used to specify the ability of the conventional, robust, and proposed models. By this purpose three different leveling networks have been simulated. The random errors ε_{i} for i=1…n were generated using a normal distribution N(0, σ^{2}) with a mean of zero and a variance of σ^{2}. Also, the good and contaminated observations were acquired by simulation technique as described in detail by Hekimoglu and Erenoglu (2007). Since the outliers are produced through simulation, it is easy to determine whether an observation is burdened before analyzing. The method is deemed successful if the observation recognized as an outlier matches the really burdened observation. The process is considered unsuccessful if it fails. When the simulated observation is chosen randomly, the successful rate indicates that the global MSR and the local MSR can be computed for each particular observation in the leveling network for 10 000 samples. The same samples were subjected to conventional, robust, and proposed methods to compare their MSRs with different scenarios. This study simulated outliers randomly chosen from small and largemagnitudes outliers (variously described as gross and influential outliers) for three leveling networks. An influential outlier is a situation that, either independently or when combined with other biased observations, adversely affects the outcomes of an analysis. Even a single influential outlier may ruin the estimation parameters. A leveling network used for the simulation has 7 points and 15 observations as seen in Fig. 2. The precision is considered to be ${\mathit{\sigma}}_{i}^{\mathrm{2}}={\mathit{\sigma}}_{\mathrm{0}}/\sqrt{S}$, where S is the length of the leveling line in kilometers and σ_{0}=1 mm ${\sqrt{\mathrm{1}\phantom{\rule{0.125em}{0ex}}\mathrm{km}}}^{\mathrm{1}}$. MSRs for 10 000 samples were calculated for each method when there were different magnitudes and different numbers of outliers in the network. The small and large outliers were generated in the intervals of [3–6σ] and [6–12σ], respectively.
As Table 1 shows, even if the number of outliers changes, the MSRs of the proposed method increase significantly compared to the conventional and robust methods. In cases where there is no outlier (e.g., m=0), the results in which the H_{0} hypothesis is rejected are also seen in Table 1. The proposed method generated typeII errors at the rate of 5 %, where the significance level was at 0.05.
Additionally, the a posteriori variance is easily influenced by outliers in the data set, which harms the abilities of methods that use the a posteriori variance. The a posteriori variance from the LSE is typically utilized as a threshold value instead of the a priori variance if the a priori variance is unknown. Therefore, the MSRs of the robust techniques of the former case are higher than the latter. As a result of these findings, only the case where the variance is known, which is less affected by an outlier, is taken into account in the results shown in the tables (Tables 2–8) to compare with the FSME hereafter.
Extensive experiments have been done to compare the proposed method with robust methods, such as the Danish and Huber methods, besides the conventional outlier detection procedures (i.e., Baarda and Pope). The redundancies are an important indicator to recognize the observations most vulnerable to bias (Durdag, 2022). The redundancy matrix is calculated from $\mathbf{R}=\mathbf{H}\mathbf{E}$, where $\mathbf{H}=\mathbf{A}({\mathbf{A}}^{\mathrm{T}}\mathbf{PA}{)}^{\mathrm{1}}{\mathbf{A}}^{\mathrm{T}}\mathbf{P}$ is a hat matrix. The local MSRs have been calculated for the specific observations with the highest and lowest redundancy in the leveling network. Among the observations, those with the two largest redundancies are h_{13} and h_{9}, and the three lowest are h_{1}, h_{7}, and h_{8}. As can be seen from the Table 4 below, MSRs increase as the redundancy does.
It is apparent from Table 2 that the highest MSRs for the biased observation h_{1} amongst the conventional and robust methods are from the Danish method (54 %). In addition, the MSR of the proposed method is higher than the Danish one by 30 %. The highest MSR has been obtained by the FSME as 94 % for the observation with the highest redundancy h_{13}. As the redundancy gets smaller, the difference in MSR between the proposed method and other methods increases.
As shown in Table 3, the highest MSRs are obtained by the FSME in contrast with other techniques for different numbers of outliers. When Tables 1 and 3 are compared, the MSRs increase with an enlargement in the magnitude of outlier.
The smearing effect of the LSE, almost equivalent to its SC (sensitivity curve), behaves systematically as a function of the partial redundancy (Durdag, 2022). For this reason, the MSRs have been calculated for the pair of observations with the lowest and largest partial redundancy with small outliers in Table 4. The neighboring observations, especially the point that has three leveling lines, are some of the most vulnerable to bias (e.g., h_{6}h_{7} and h_{8}h_{7}) in the leveling network (Fig. 2). The local MSRs are lower than the global MSRs, in the case m=2 in Table 1, for observations most vulnerable to bias with both the lowest and highest redundancies as shown in Table 4.
The results, as shown in Table 4, indicate that the MSRs of observation h_{6}h_{7} with the highest partial redundancies increases compared to h_{13}h_{1} with the lowest ones by almost 30 % for the Baarda and Danish models and 50 % for the FSME model.
It is apparent from Table 5 that Baarda and Danish are the two most successful methods against gross and influential outliers among classical and robust methods. In the case of small outliers with gross or influential outliers, the robustness of the models has been tested. Different types of outliers have been generated to evaluate the MSRs of the models for various scenarios as follows: (I) gross outlier (50σ), (II) influential outliers (1000σ), (III) a small outlier and a gross outlier, (IV) a small outlier and an influential outlier, (V) two small outliers and a gross outlier, and (VI) two small outliers and an influential outlier.
Comparing Table 5 with Table 6, it is observable how MSRs of these two methods were affected in case of one or two small outliers. If a small outlier occurs, the MSRs drop dramatically by about 40 %. The MSRs drop to 20 % with two small outliers in the network. Furthermore, this loss is around 35 %–40 % for the proposed method. The FSME, however, stands out as the model with an MSR of 60 %–70 % in scenarios involving small outliers.
When the redundancies of the observations decrease in the leveling network, difficulties arise in determining the outliers due to the swamping and masking effects. Two different leveling networks are considered to obtain the MSR of the methods in such cases. In the first of these, the MSR of the models has been compared by excluding an observation of the network. As seen in Table 7, the MSR decreased by 30 % in all models when m=2, compared with the case m=1. Although the number of small outliers changes, the highest MSRs have been obtained by the FSME for the leveling network 2a. The network is further weakened, so only two lines of the corner point P.5 remain in the leveling network 2b.
The results, as shown in Table 8, indicate that the FSME is the model with the highest MSR for m=1. When m>1 is compared with the case m=1 in Table 8, MSRs of the conventional and robust methods show a more dramatic decrease than the FSME model. Comparing the estimated results for the networks 2a and 2b reveals an approximately 15 % drop in MSR values when m=1. MSRs decrease as the controllability of the observations in the network decreases.
Since geodetic observations are utilized in studies requiring high accuracy for determining deformations, detecting and identifying outliers become increasingly critical. Researchers commonly favor conventional and robust methods based on the additive bias model. However, this study contributes to our understanding by advancing the modeling of outliers as an additional unknown parameter within the Gauss–Markov model. The aim of this study was to evaluate the suitability of the FSME method within geodetic networks. To achieve this objective, the FSME method was applied to a leveling network. The design of the FSME method is based on identifying the minimum variance from all possible combinations that assume observations as model errors in the Gauss–Markov model. Although only a leveling network has been simulated, the functional and stochastic models of the FSME methods can be applied to all type of geodetic networks. This model yields more reliable results by preventing the swamping and masking effect. The MSRs of the suggested method were obtained for various outliers in three different leveling networks. The results of this investigation show that the FSME is a more efficient model than the robust and conventional methods. Specifically, the proposed method enhanced the MSR by almost 40 %–45 % compared to the Baarda and Danish (with the variance unknown case) methods for multiple outliers (i.e., 1<m<4). Moreover, in scenarios where specific observations were absent at corner points in leveling network 1, the proposed method exhibited 20 %–30 % greater success than alternative methods. Despite the proposed model demonstrating a higher MSR than other methods, the FSME method may encounter numerous combinations depending on the presence of observations and outliers. To address this challenge, particularly in realworld applications, the MSS (maximum subsample method) proposed by Neitzel (2004) and Ebeling (2014) offers a promising approach to reduce the number of combinations in outlier detection procedures. As demonstrated by Ebeling in deformation monitoring, MSS holds potential as a valuable tool to enhance the applicability of the proposed method, particularly within extensive geodetic networks.
The current code version is available from the project website: https://doi.org/10.5281/zenodo.10417506 (Durdağ, 2023) under the MIT license. The exact version of the model used to produce the results used in this paper is archived on Zenodo (https://doi.org/10.5281/zenodo.10417506, Durdağ, 2023), as are input data and scripts to run the model and produce the plots for all the simulations presented in this paper.
The author has declared that there are no competing interests.
Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors.
The author expresses sincere gratitude to the anonymous referees and Xinyue Yang for their valuable comments and suggestions, which have significantly enhanced the quality of this paper.
This paper was edited by Yongze Song and reviewed by two anonymous referees.
Aydin, C.: Power of global test in deformation analysis, J. Surv. Eng., 138, 51–56, https://doi.org/10.1061/(ASCE)SU.19435428.0000064, 2012.
Baarda, W.: A testing procedure for use in geodetic networks. Publications on Geodesy, 2 = 5, Netherlands Geodetic Commission, Delft, the Netherlands, ISBN 90 6132 209 X, 1968.
Batilović, M., Sušić, Z., Kanović, Ž., Marković, M. Z., Vasić, D., and Bulatović, V.: Increasing efficiency of the robust deformation analysis methods using genetic algorithm and generalised particle swarm optimisation, Surv. Rev., 53, 193–205, https://doi.org/10.1080/00396265.2019.1706294, 2021.
Duchnowski, R.: Sensitivity of robust estimators applied in strategy for testing stability of reference points. EIF approach, Geodesy and Cartography, 60, 123–134, https://doi.org/10.2478/v102770120011z, 2011.
Durdağ, U. M.: Godesist/OutlierDetectionForGeodeticLevelingNetwork: Initial Release (0.1.0), Zenodo [code and data set], https://doi.org/10.5281/zenodo.10417506, 2023.
Durdag, U. M., Hekimoglu, S., and Erdogan, B.: What is the relation between smearing effect of least squares estimation and its influence function? Surv. Rev., 54, 320–331, https://doi.org/10.1080/00396265.2021.1939590, 2022.
Ebeling, A.: GroundBased Deformation Monitoring. PhD Thesis, University of Calgary, Department of Geomatics Engineering, Calgary, https://doi.org/10.11575/PRISM/26325, 2014.
Erdogan, B.: An outlier detection method in geodetic networks based on the original observations, Bol. Ciênc. Geod., 20, 578–589, https://doi.org/10.1590/S198221702014000300033, 2014.
Erdogan, B., Hekimoglu, S., Durdag, U. M., and Ocalan, T.: Empirical estimation of the power of test in outlier detection problem, Studia Geophys. et Geod., 63, 55–70, https://doi.org/10.1007/s1120001811449, 2019.
Gao, Y., Krakiwsky, E. J., and Czompo, J.: Robust testing procedure for detection of multiple blunders, J. Surv. Eng., 118, 11–23, https://doi.org/10.1061/(ASCE)07339453(1992)118:1(11), 1992.
Hekimoglu, S.: Reliabilities of χ^{2}and F Tests in GaussMarkov Model, J. Surv. Eng., 125, 109–135, https://doi.org/10.1061/(ASCE)07339453(1999)125:3(109), 1999.
Hekimoglu, S.: Increasing reliability of the test for outliers whose magnitude is small, Surv. Rev., 38, 274–285, https://doi.org/10.1179/sre.2005.38.298.274, 2005.
Hekimoglu, S. and Erenoglu, R. C.: Effect of heteroscedasticity and heterogeneousness on outlier detection for geodetic networks, J. Geod., 81, 137–148, https://doi.org/10.1007/s001900060095z, 2007.
Hekimoglu, S. and Koch, K. R.: How can reliability of the test for outliers be measured?, Allgemeine Vermessungsnachrichten, 107, 247–254, 2000.
Hekimoglu, S., Erdogan, B., and Butterworth, S.: Increasing the efficacy of the conventional deformation analysis methods: alternative strategy, J. Surv. Eng., 136, 53–62, https://doi.org/10.1061/(ASCE)SU.19435428.0000018, 2010.
Hekimoglu, S., Erdogan, B., Soycan, M., and Durdag, U. M.: Univariate approach for detecting outliers in geodetic networks, J. Surv. Eng., 140, 04014006, https://doi.org/10.1061/(ASCE)SU.19435428.0000123, 2014.
Hekimoglu, S., Erdogan, B., and Erenoglu, R. C.: A new outlier detection method considering outliers as model errors, Exp Techniques, 39, 57–68, https://doi.org/10.1111/j.17471567.2012.00876.x, 2015.
Huber, P. J.: Robust Estimation of a Location Parameter, Ann. Math. Stat., 35, 73–101, http://www.jstor.org/stable/2238020 (last access: 12 March 2024), 1964.
Huber, P. J.: Robust Statistics, John Wiley and Sons, New York, USA, 312 pp., ISBN 9780471418054, 1981.
Koch, K. R.: Parameter estimation and hypothesis testing in linear models, Springer, New York, USA, 334 pp., ISBN 9783540652571, 1999.
Lehmann, R.: 3σrule for outlier detection from the viewpoint of geodetic adjustment, J. Surv. Eng., 139, 157–165, https://doi.org/10.1061/(ASCE)SU.19435428.0000112, 2013.
Maronna, R. A., Martin, R. D., Yohai, V. J., and SalibiánBarrera, M.: Robust statistics: theory and methods (with R), Wiley, Chichester, UK, 430 pp., ISBN 9781119214687, 2019.
Neitzel, F.: Identifizierung konsistenter Datengruppen am Beispiel der Kongruenzuntersuchung geodätischer Netze, PhD thesis, Deutsche Geodätische Kommission, Reihe C, Nr. 565, München, ISSN 0065 5325, ISBN 3 7696 5004 2, https://www.researchgate.net/profile/FrankNeitzel/publication/35226424 (last access: 14 March 2024), 2004.
Niemeier, W.: Anlage von Überwachungsnetzen. Geodaetische Netze in Landes und Ingenieurvermessung II., edited by: Pelzer, H., Verlag Konrad Wittwer, Stuttgart, Germany, 527–558, ISBN 3879191298, LCCN 80491066, 1985 (in German).
Nowel, K.: Specification of deformation congruence models using combinatorial iterative DIA testing procedure, J. Geod., 94, 118, https://doi.org/10.1007/s00190020014469, 2020.
Pope, A. J.: The statistics of residuals and the detection of outliers. NOAA Technical Rep. NOS 65 NGS 1, U.S. Dept. of Commerce, Rockville, Maryland, USA, 98 pp., https://repository.library.noaa.gov/view/noaa/30811 (last access: 13 March 2024), 1976.
Teunissen, P. J.: Testing theory; an introduction, VSSD Leeghwaterstraat 42, 2628 CA Delft, the Netherlands, 147 pp., ISBN 9789040719752, 2006.
Wang, J., Zhao, J., Liu, Z., and Kang, Z.: Location and estimation of multiple outliers in weighted total least squares, Measurement, 181, 109591, https://doi.org/10.1016/j.measurement.2021.109591, 2021.
Zienkiewicz, M. H. and Dąbrowski, P. S.: Matrix Strengthening the Identification of Observations with Split Functional Models in the Squared M_{split}(q) Estimation Process, Measurement, 217, 112950, https://doi.org/10.1016/j.measurement.2023.112950, 2023.