1 Introduction

Absolute and relative control networks are distinguished in deformation measurements and analysis. Absolute control networks consist of controlled points placed on the surveyed object (e.g., on a dam, bridge) and potential reference points usually placed outside the effect of factors causing deformations. Relative control networks consist only of controlled points established on the surveyed object (e.g., on an area covered by mining damage). Both potential reference points of absolute networks and potential stable points of relative networks (the points are on a potentially rigid network fragment) form FCN (free control networks). The problem of deformation analysis for FCN is, in fact, the problem of identifying the mutual stability of these network points. In this way, the datum for the displacement vector of all absolute or relative control network points is defined. It can be seen that deformation analysis for FCN is, therefore, an essential task in deformation analysis for control networks (Chen 1983; Caspary 2000; Denli 2008).

Methods of deformation analysis for FCN described in literature can be classified according to different criteria. In this paper, they are divided into robust methods and non-robust methods (Chrzanowski and Chen 1990; Caspary 2000; Setan and Singh 2001). Today, the robust methods are applied more often. It should be stressed that “robust” does not refer here to outliers in observations; “robust” refers here to single-point movements. The observations must be clean of outliers. This is very important. Classical robust methods are based on the results of separate adjustments by the least squares (LS) method of two measurement epochs, i.e., on adjusted coordinates. These methods consist in S-transformation (Helmert similarity transformation) of differences in these coordinates with the optimization condition of robust estimation. The obtained displacement vectors of single points are assessed statistically for significance. The best-known methods from this group include the iterative weighted similarity transformation (IWST) method, which fulfills the condition of the minimum sum of absolute component values of displacement vectors (Chen 1983; Chen et al. 1990; Setan and Singh 2001; Setan and Othman 2006; Gökalp and Tasçi 2009) and the least absolute sum (LAS) method, which fulfills the condition of the minimum sum of length values of displacement vectors (Caspary 1984; Caspary and Borutta 1987; Caspary et al. 1990; Setan and Singh 2001). Proposals for the use of other optimization conditions, e.g., the minimum objective function of the Huber method or the Danish method, can also be found in literature (Caspary and Borutta 1987). The IWST method serves as the prototype for classical robust methods. The other robust methods are only modifications, differing in the optimization condition used. The non-robust methods are also based on the results of separate adjustments by the LS method of two measurement epochs, i.e., on adjusted coordinates. Statistical tests are most popular in this group, e.g., the congruency test (Niemeier 1981; Setan and Singh 2001; Denli and Deniz 2003). The congruency test method consists in finding the most numerous possible group of points for which the value of the displacement test statistic does not exceed the critical value. Displacement vectors are computed on the basis of S-transformation of differences in adjusted coordinates of these points with the optimization condition of the LS method or on the basis of differences in coordinates obtained from free adjustments by the LS method. Although many more other new methods for deformation analysis, e.g., robust method based on \(\hbox {M}_{\mathrm{split}}\) estimation (Wiśniewski 2009; Duchnowski and Wiśniewski 2012) or R estimation (Duchnowski 2010, 2013) or the non-robust Fredericton approach (Gökalp and Tasçi 2009), can be found in literature, this paper is limited to the presentation of only the most common solutions.

Geodetic methods are most often used to measure minor deformations with values slightly exceeding errors in their measurement. A special role in deformation analysis is, therefore, played by the quality of results, i.e., the accuracy of the determined displacement vectors and the correct assessment of their statistical significance (effectiveness). The quality of deformation analysis results is conditioned by the measurement precision (random errors) and the measurement reliability and the processing of the results. A higher reliability level of a measurement produces a lower risk of making gross and systematic errors. A high measurement reliability level is achieved by, among others, periodical instrument checks, the use of self-checking measurement procedures and the introduction of corrections due to the disturbing effects of external conditions. The higher the reliability level of the processing of measurement results, the higher the probability will be of detecting possible gross and systematic errors (internal reliability) and the lower the effect of these undetected errors will be on the quality of deformation analysis results (external reliability), e.g., Caspary (2000), Prószyński (2010).

Deformation measurements are performed at different moments of time, but often with the same equipment, methods, geometric conditions and in a similar environment in epochs 1 and 2. Moreover, the same instrument and target are most often set on given points in both epochs. It is, therefore, reasonable to assume that the results of deformation measurements can be distorted both by random errors and by some non-random errors, which are constant in both epochs, i.e., errors with the same value and the same sign in both epochs. This means that the error of measurement of the same geometric element of the network in both epochs may contain a common factor, which is constant in both epochs. This factor is, in fact, the difference in the measurement accuracy and precision. The constant errors are understood, but the manifestation of these errors is difficult to determine in practice. Classical deformation analysis of FCN is most often performed by robust methods. In these methods, the displacement vector is determined in robust S-transformation of differences in adjusted coordinates. This approach does not completely eliminate the effect of constant errors and, therefore, leads to reduced external reliability level of processing and, consequently, to reduced quality of deformation analysis results. Moreover, the displacement vector and its covariance matrix are estimated separately in two different computational processes. As a result, the displacement vector estimator and its covariance matrix estimator have different stochastic properties, i.e., the actual and estimated accuracy of the displacement vector estimator is different.

In the 1970s, the Polish geodesist Prof. Tadeusz Lazzarini dealt with this problem for control networks with a defined datum. Prof. Lazzarini developed a method for joint estimation of the displacement vector and its covariance matrix directly from differences in unadjusted observations (Lazzarini et al. 1977; Chrzanowski and Chen 1990). This approach completely solves, among others, the problem of the effect of constant errors and the problem of discrepancy between the actual and estimated accuracy of the displacement vector estimator. This method is still used successfully in deformation analysis of control networks with a defined datum, among others in an automated system for continuous control measurements ALERT (Wilkins et al. 2003). Modern robust estimation and matrix g-inverse theory allow this approach to be expanded to FCN (control networks without a defined datum). Such expansion is the aim of this paper. In fact, a new robust method for FCN deformation analysis will be developed and examined in this paper, in which the displacement vector will be determined in the process of robust M estimation of differences in unadjusted observations. The proposed method was conventionally called robust estimation of deformation from observation differences (REDOD).

2 Robust transformation of deformation from coordinate differences

The robust FCN deformation analysis method proposed in this paper will be developed based on classical robust methods and is specifically based on the IWST method. Therefore, the algorithm of classical robust methods was derived and described in this section. Particular attention was paid to the IWST method.

As previously mentioned, in classical methods the displacement vector is determined in the process of robust S-transformation of the differences in adjusted coordinates. Generally, the algorithm always consists of the following three steps.

2.1 Step 1: LS estimation of the vector of coordinates

LS estimation of the vector of coordinates is carried out separately for each measurement epoch. The computations cannot distort measurement results or the relationships occurring between them. The basic aim is to provide data for starting robust transformation. The result of this processing is the vector of adjusted coordinates for all analyzed points in an FCN

$$\begin{aligned} {\hat{\mathbf{x}}}=\mathbf{x}^{0}+{\hat{{\varvec{\updelta }}}}_\mathbf{x}, \end{aligned}$$
(1)

where \(\mathbf{x}^{0}\in \mathfrak {R}^{u\times 1}\) is the vector of the approximate coordinates (the same in both epochs) and \({\hat{{\varvec{\updelta }}}}_\mathbf{x} \in \mathfrak {R}^{u\times 1}\) is the LS estimator of the vector of the corrections to the approximate coordinates.

The functional model, which is the linearized form of the initial nonlinear relationships, has the form

$$\begin{aligned} \mathbf{l}^\mathrm{obs}-\mathbf{l}^{0}+\mathbf{v}={\mathbf{A}{\varvec{\updelta }}}_\mathbf{x}, \quad \mathbf{l}^\mathrm{obs}\sim N\left( {\mathbf{Ax},\mathbf{C}} \right) , \end{aligned}$$
(2)

where \(\mathbf{A}\in \mathfrak {R}^{n\times u}\) is the design matrix, \(\mathbf{l}^{0}\in \mathfrak {R}^{n\times 1}\) is the vector of approximate observations, \(\mathbf{l}^\mathrm{obs}\in \mathfrak {R}^{n\times 1}\) is the vector of actual observations, \(\mathbf{v}\in \mathfrak {R}^{n\times 1}\) is the vector of residuals of observations and \(\mathbf{C}\in \mathfrak {R}^{n\times n}\) is the covariance matrix of observations. The sought parameter vector estimator has the form

$$\begin{aligned} {\hat{{\varvec{\updelta }}}}_\mathbf{x} =\mathbf{Q}_{{\hat{\mathbf{x}}}} {\varvec{{\upomega }}}, \end{aligned}$$
(3)

where \({\varvec{{\upomega }}}=\mathbf{A}^\mathrm{T}\mathbf{P}(\mathbf{l}^\mathrm{obs}-\mathbf{l}^{0}),\,\mathbf{P}\in \mathfrak {R}^{n\times n}\) is the weight matrix of observations, \(\mathbf{Q}_{{\hat{\mathbf{x}}}} =\mathbf{N}^{-}\) is the cofactor matrix of the vector of adjusted coordinates and \(\mathbf{N}^{-}=(\mathbf{A}^\mathrm{T}\mathbf{PA})^{-}\) is any g-inverse, fulfilling the property \(\mathbf{NN}^{-}\mathbf{N}=\mathbf{N}\). The network is a free network. In this case, the network has a datum defect

$$\begin{aligned} de=u-\hbox {rank}(\mathbf{A}), \quad \hbox {rank}(\mathbf{A})=r<u \end{aligned}$$
(4)

the matrix A is the matrix of the columnarly incomplete rank and, therefore, the ordinary (classical) inverse \(\mathbf{N}^{-1}\) will not be determined. The g-inverse determination methods are well-described in literature (e.g., Rao 1973; Caspary 2000). For example, one of the g-inverses can be determined in a simple, algebraic manner, by representation of the design matrix in the following form

$$\begin{aligned} \mathbf{A}=\left[ {{\begin{array}{l@{\quad }l} {\mathbf{A}_1}&{} {\mathbf{A}_2} \\ \end{array}}} \right] , \end{aligned}$$
(5)

where \(\mathbf{A}_1 \in \mathfrak {R}^{n\times r},\,r=\hbox {rank}(\mathbf{A}_1)=\hbox {rank}(\mathbf{A})\) and \(\mathbf{A}_2 \in \mathfrak {R}^{n\times de}\) is, in fact, the matrix defining the datum for the geodetic network. The final result of step 1 is the vectors \({\hat{\mathbf{x}}}_{(e)}\) (1) and their cofactor matrices \(\mathbf{Q}_{{\hat{\mathbf{x}}}_{(e)}}\), where \(e\) is the number of the measurement epoch.

Of course, during LS estimation of the vector of coordinates, other important aspects that need to be considered are outlier detection (observations) (e.g., Baarda’s data snooping, Pope’s test or some robust estimation) and tests of the compatibility between the two epochs (variance ratio test). Moreover, in the case of heterogeneous networks, it is sometimes important to perform the estimation of the variance and covariance components.

2.2 Step 2: robust S-transformation of the displacement vector

Displacements of individual points in an FCN cannot be determined on the basis of differences in the direct values of the vectors \({\hat{\mathbf{x}}}_{(1)} \) and \({\hat{\mathbf{x}}}_{(2)} \). LS estimation of the vector of coordinates can concern networks with different datum defect types in both epochs (because of different types of observations). Moreover, constraints defining the datum for the geodetic network in epochs 1 and 2 can concern different points (e.g., because of damage to some points in epoch 2) or can concern the same unstable points. For this reason, the datum for the geodetic network in epoch 2 can be freely shifted, rotated and re-scaled in relation to the datum in epoch 1 (Chen 1983; Chen et al. 1990). The vector \({\hat{\mathbf{x}}}_{(2)} \) can, therefore, be adjusted in a different datum than the vector \({\hat{\mathbf{x}}}_{(1)}\) and this is why

$$\begin{aligned} \mathbf{d}\ne {\hat{\mathbf{x}}}_{(2)} -{\hat{\mathbf{x}}}_{(1)}, \end{aligned}$$
(6)

where \(\mathbf{d}\in \mathfrak {R}^{u\times 1}\) is the vector of displacement components for all analyzed points in an FCN. This difference \({\hat{\mathbf{x}}}_{(2)} -{\hat{\mathbf{x}}}_{(1)} \) is the so-called apparent (or raw) displacement vector.

To reduce the vectors \({\hat{\mathbf{x}}}_{(1)} \) and \({\hat{\mathbf{x}}}_{(2)} \) to a common datum, the vector \({\hat{\mathbf{x}}}_{(2)} \) is transformed to the datum of the vector \({\hat{\mathbf{x}}}_{(1)} \). This task consists in fitting the vector \({\hat{\mathbf{x}}}_{(2)} \) into the vector \({\hat{\mathbf{x}}}_{(1)} \) reduced to the center of gravity of the network. The transformed vector \({\hat{\mathbf{x}}}_{(2)}^{\prime } \) has the form

$$\begin{aligned} {\hat{\mathbf{x}}}_{(2)}^{\prime } ={\hat{\mathbf{x}}}_{(2)} -\mathbf{Ht} \end{aligned}$$
(7)

because \({\hat{\mathbf{x}}}_{(2)} ={\hat{\mathbf{x}}}_{(2)}^{\prime } +\mathbf{Ht}\), where \(\mathbf{t}\in \mathfrak {R}^{h\times 1}\) is the vector of S-transformation parameters, i.e., translation, rotation, scale distortion, \(h\) designates the number of all datum parameters of a given network and H is the design matrix of S-transformation to a common datum (epoch 1). For example, for a 2-D network, the matrix H has the form

$$\begin{aligned} \mathbf{H}\in \mathfrak {R}^{2m\times h}=\left[ {{\begin{array}{cccc} \vdots &{} \vdots &{} \vdots &{} \vdots \\ 1&{} 0&{} {-\bar{{y}}_{i(1)}} &{} {\bar{{x}}_{i(1)}} \\ 0&{} 1&{} {\bar{{x}}_{i(1)}} &{} {\bar{{y}}_{i(1)}} \\ \vdots &{} \vdots &{} \vdots &{} \vdots \\ \end{array}}}\right] , \end{aligned}$$
(8)

where \(m\) designates the number of points in a given network and \(\bar{{x}}_{i(1)},\,\bar{{y}}_{i(1)} \) designate adjusted coordinates of point \(i\) in epoch 1 reduced to the center of gravity of the network. The first two columns refer to translation along the axes \(X, Y\), respectively, the third column refers to rotation around the axis \(Z\) and the last column refers to scale distortion. More theory on that issue (among others, the matrix H for a 1-D and 3-D network) can be found in papers by Chen (1983), Caspary (2000) and Even-Tzur (2011).

Since the displacement vector now has the form \(\mathbf{d}={\hat{\mathbf{x}}}_{(2)}^{\prime } -{\hat{\mathbf{x}}}_{(1)} \), consequently \(\mathbf{d}=({\hat{\mathbf{x}}}_{(2)} -{\hat{\mathbf{x}}}_{(1)} )-\mathbf{Ht}\). Of course, this system of equations has infinitely many solutions for the displacement vector. To obtain a unique, sought value of this vector, we also need to formulate proper objective functions.

Let us recall that FCN may consist of potential reference points of absolute networks or potential stable points of relative networks. Hence, we can assume that most of these points (more than half) are actually stable. Unfortunately, there is a risk that single unstable points may occur among this group. It can be assumed that measurement errors at all epochs are a sample of random variables with a normal distribution with a mean of zero and accepted standard deviation. In this case, the estimated displacements of actually stable points should also be a sample of random variables with such a distribution and the estimated displacements of actually unstable points should be random variables which are the outliers of this distribution. In other words, the vector of displacement components of FCN points may have a contaminated normal distribution (analogously as, e.g., a vector of observations with suspected gross errors). This is why the robust objective functions are very useful in deformation analysis for FCN. Thus, “outliers” do not refer here to the observations. “Outliers” refer here to the single-point movements. It should be stressed that if all of the FCN points are certainly actually stable, then the LS and robust methods lead to unbiased results, but robust methods are less efficient and yield greater variances. This decrease in efficiency is the price to be paid for the robustness of the method.

In the IWST method, the objective function is the \(\hbox {L}_{1}\)-norm of vector \(\mathbf{d}\), i.e., (Chen 1983; Chen et al. 1990; Setan and Singh 2001; Setan and Othman 2006; Gökalp and Tasçi 2009). The optimization problem of the IWST method has the following form

(9)

where \(\mathbf{d}=[d_{1},\,\ldots ,\,d_u]^\mathrm{T},\,d_i\) is the given component of vector \(\mathbf{d}\) for point \(i\) (\(dx_i\) or \(dy_i\) or \(dz_i\)), \(u\) is the number of the displacement components for all \(m\) analyzed points in an FCN. The final solution to the optimization problem of the IWST method has the following iterative form

$$\begin{aligned} \left. {\begin{array}{c} {\hat{\mathbf{d}}}^{(k)}=\mathbf{S}^{(k)}\left( {{\hat{\mathbf{x}}}_{(2)} -{\hat{\mathbf{x}}}_{(1)}} \right) \\ \mathbf{Q}_{{\hat{\mathbf{d}}}}^{(k)} =\mathbf{S}^{(k)}\left( {\mathbf{Q}{}_{{\hat{\mathbf{x}}}_{(1)}} + \mathbf{Q}_{{\hat{\mathbf{x}}}_{(2)}}} \right) \left( {\mathbf{S}^{(k)}} \right) ^\mathrm{T} \\ \mathbf{W}^{(k+1)}=\hbox {diag}\left( {\ldots ,\,w_i^{(k+1)},\, \ldots }\right) \end{array}} \right\} _{k=1,2,\ldots } \end{aligned}$$
(10)

where \(\mathbf{S}^{(k)}=\mathbf{I}-\mathbf{H}(\mathbf{H}^\mathrm{T}\mathbf{W}^{(k)}\mathbf{H})^{-1}\mathbf{H}^\mathrm{T}\mathbf{W}^{(k)}\) is the matrix of S-transformation of the displacement vector to a common datum in epoch 1, \(\mathbf{W}^{(k=1)}=\mathbf{I},\,\mathbf{I}\in \mathfrak {R}^{u\times u}\) is the identity matrix and

$$\begin{aligned} w_i^{(k+1)} =1/\left| {\hat{{d}}_i^{(k)}} \right| \end{aligned}$$
(11)

It is possible that during the iterations (10) some \(\hat{{d}}_i^{(k)} \) may approach zero, causing numerical instabilities, because \(1/\left| {\hat{{d}}_i^{(k)}} \right| \) becomes very large. For this reason in practice, Eq. (11) is replaced with the equation

$$\begin{aligned} w_i^{(k+1)} =1\big /\left( {\left| {\hat{{d}}_i^{(k)}} \right| +c} \right) , \end{aligned}$$
(12)

where \(c\) is the precision for computations, e.g., 0.0001 m. Proposals for uses of other solutions in this task can be found in other papers, e.g., Marx (2013). The iterative process (10) is finished when all components of the difference \(\left| {{\hat{\mathbf{d}}}^{(k+1)}-{\hat{\mathbf{d}}}^{(k)}} \right| \) are lower than the adopted precision for computations \(c\) (convergence). The final robust estimator of the vector of displacement components for all analyzed points in an FCN, i.e., vector \({\hat{\mathbf{d}}}^{R}\) and the final cofactor matrix of this vector, i.e., the matrix \(\mathbf{Q}_{{\hat{\mathbf{d}}}^{R}} \), is obtained from the last iterative step. Vector \({\hat{\mathbf{d}}}^{R}\) consists of vectors \({\hat{\mathbf{d}}}_i^R \) for single points in an FCN, i.e., \(({\hat{\mathbf{d}}}^{R})^\mathrm{T}=[\ldots ,({\hat{\mathbf{d}}}_i^R )^\mathrm{T},\ldots ]\), and the matrix \(\mathbf{Q}_{{\hat{\mathbf{d}}}^{R}} \), neglecting the correlation between the points and contains cofactor matrices of these vectors, i.e., \(\mathbf{Q}_{{\hat{\mathbf{d}}}^{R}} =\hbox {diag}(\ldots ,\mathbf{Q}_{{\hat{\mathbf{d}}}_i^R},\ldots )\). Proposals for uses of other objective functions \(\psi (\mathbf{t})\) in this task (among others, the objective function of the LAS method, the Huber method or the Danish method) can be found in papers by Caspary and Borutta (1987), Caspary et al. (1990), Setan and Singh (2001).

2.3 Step 3: assessment of the significance of the displacement vector

All \({\hat{\mathbf{d}}}_i^R \) vectors are assessed statistically for significance. The aim of this assessment is to check if these vectors are actually displacements or only the result of random measurement errors. The first step consists in checking the global significance test

$$\begin{aligned} T=\frac{\left( {{\hat{\mathbf{d}}}^{R}} \right) ^\mathrm{T}\mathbf{Q}_{{\hat{\mathbf{d}}}^{R}}^+ \left( {{\hat{\mathbf{d}}}^{R}} \right) }{u\hat{{\sigma }}_0^2} <F\left( {\alpha ,u,f} \right) , \end{aligned}$$
(13)

where

$$\begin{aligned} \hat{{\sigma }}_0^2 =\left( {{\hat{\mathbf{v}}}_{(1)}^\mathrm{T} \mathbf{P}_{(1)} {\hat{\mathbf{v}}}_{(1)} +{\hat{\mathbf{v}}}_{(2)}^\mathrm{T} \mathbf{P}_{(2)} {\hat{\mathbf{v}}}_{(2)}} \right) \big /f \end{aligned}$$
(14)

is the pooled variance factor estimator for both epochs and \({\hat{\mathbf{v}}}_{(e)} ={\mathbf{A}\hat{{\varvec{\updelta }}}}_{\mathbf{x}_{(e)}} -(\mathbf{l}_{(e)}^\mathrm{obs} -\mathbf{l}^{0}),\,f=f_{(1)} +f_{(2)} ,\,f_{(e)} =n_{(e)} -u_{(e)} +de_{(e)} ,\,\alpha \) is the significance level, i.e., the probability of making a Type I error in testing statistical, \(u=\hbox {rank}(\mathbf{Q}_{{\hat{\mathbf{d}}}^{R}} )\) and \(F\) is the critical value read from Fisher’s cumulative distribution function (Aydin 2013). If the global test passes, this means that all \({\hat{\mathbf{d}}}_i^R \) vectors are not statistically significant and, thus, these points can be regarded as actually stable. In the opposite case, localization of single-point displacements is then performed. The single-point test, neglecting the correlation between the points, is

$$\begin{aligned} T_i =\frac{\left( {{\hat{\mathbf{d}}}_i^R} \right) ^\mathrm{T}\mathbf{Q}_{{\hat{\mathbf{d}}}_i^R}^{-1} \left( {{\hat{\mathbf{d}}}_i^R} \right) }{u_i \hat{{\sigma }}_0^2} <F\left( {\alpha ,u_i ,f} \right) , \end{aligned}$$
(15)

where \(u_i =\hbox {rank}(\mathbf{Q}_{{\hat{\mathbf{d}}}_i^R} )\) is the number of the dimension of the single-point displacement vector (1 or 2 or 3) (Chen 1983; Setan and Singh 2001). Instead of the local test (15), neglecting the correlation between the points, one can also apply the local test based on the decorrelated matrix \(\mathbf{Q}_{{\hat{\mathbf{d}}}^{R}} \). More details are given in Caspary (2000). If the single-point test passes, this means that the displacement of point \(i\) is not statistically significant and, thus, this point can be regarded as actually stable. In the opposite case, this point can be regarded as unstable. The solution (15) can be used for 1-D, 2-D and 3-D networks. Moreover, another condition can be derived from the condition (15) for better interpretation of results

$$\begin{aligned} {\hat{\mathbf{d}}}_i^R \in E=f\left( {F,\hat{{\sigma }}_0^{2}, \mathbf{Q}_{{\hat{\mathbf{d}}}_i^R}} \right) , \end{aligned}$$
(16)

where \(E\) is the confidence interval (1-D network) or the set of points formed by the confidence ellipse (2-D network), (e.g., Kamiński and Nowel 2013) or the confidence ellipsoid (3-D network) (e.g., Cederholm 2003). In this case, the significance assessment consists in checking graphically whether the vector \({\hat{\mathbf{d}}}_i^R \) does not exceed the confidence interval for determination of this vector (1-D network) or the confidence ellipse for determination of this vector (2-D network) or the confidence ellipsoid for determination of this vector (3-D network). Of course, both solutions (15), (16) give the same results.

It should be stressed that the matrix \(\mathbf{Q}_{{\hat{\mathbf{d}}}^{R}} \) is approximate since the relation between \({\hat{\mathbf{x}}}_{(2)} -{\hat{\mathbf{x}}}_{(1)} \) and \({\hat{\mathbf{d}}}^{R}\) is nonlinear according to (10). Thus, the law of variance propagation does not apply. In this case, the proper cofactor matrix can, of course, be determined only by the Monte Carlo approach (e.g., Du Mond and Lenth 1987). Moreover, the assumption that the \(T_i \) statistic has Fisher’s distribution is also approximate based on the assumption \({\hat{\mathbf{d}}}^{R}\sim {N}\). The relation between \(({\hat{\mathbf{x}}}_{(2)} -{\hat{\mathbf{x}}}_{(1)} )\sim {N}\) and \({\hat{\mathbf{d}}}^{R}\) is nonlinear and, therefore, the estimator \({\hat{\mathbf{d}}}^{R}\) does not have a normal distribution. The distribution of this estimator is not known and the initial assumption \({\hat{\mathbf{d}}}^{R}\sim {N}\) is only approximate. Consequently, the assumptions \(({\hat{\mathbf{d}}}_i^R )^\mathrm{T}\mathbf{Q}_{{\hat{\mathbf{d}}}_i^R}^{-1} ({\hat{\mathbf{d}}}_i^R )\sim \upchi ^{2}\) and \(T_i \sim {F}\) are also approximate, although the approximate \(F\) tests (13, 15, 16) are accepted.

3 Robust estimation of deformation from observation differences

As previously mentioned, classical robust methods for deformation analysis of FCN are based on the results of separate adjustments of observations for epochs 1 and 2, i.e.,

The basis for such an approach is the assumption that vector d is the difference in functions of independent observations \((\mathbf{d}=f(\mathbf{l}_{(2)}^\mathrm{obs} )-f(\mathbf{l}_{(1)}^\mathrm{obs}))\). If the observations are distorted in both epochs only by random errors, then such an approach is correct, but not if observations are additionally distorted by constant errors.

First (and most importantly), separate adjustments fulfill the condition for the closing of geometric figures, which is not significant in the deformation analysis process. Such a procedure, in the case of the occurrence of constant errors, leads to an unjustified increase in the values of the vectors \({\hat{\mathbf{v}}}_{(1)},\,{\hat{\mathbf{v}}}_{(2)} \) and, thus, the factor \(\hat{{\sigma }}_0^2 \) (14) and, consequently, may lead to incorrect assessment of the statistical significance of the \({\hat{\mathbf{d}}}_i^R \) vectors (13), (15) or (16). Second, if the matrices \(\mathbf{P}_{(1)},\,\mathbf{P}_{(2)} \) are not proportional in both epochs (e.g., for GNSS vectors), vector \({\hat{\mathbf{d}}}^{R}\) will be distorted by the effect of constant errors. These errors will not cancel out completely in the iterative process (10) because, in this case, their effect on the values of the vectors \({\hat{\mathbf{x}}}_{(1)},\,{\hat{\mathbf{x}}}_{(2)} \) (1) will be different. As a result, the vector \({\hat{\mathbf{d}}}^{R}\) will be partly distorted by their effect. Third, the vector \({\hat{\mathbf{d}}}^{R}\) and the factor \(\hat{{\sigma }}_0^2 \) are estimated independently, in two different computational processes. As a result, the vector \({\hat{\mathbf{d}}}^{R}\) and the factor \(\hat{{\sigma }}_0^2 \) have different stochastic properties. This discrepancy may lead to incorrect assessment of the statistical significance of the vectors \({\hat{\mathbf{d}}}_i^R \).

As mentioned in the introduction, when the displacement vector is determined directly from differences in unadjusted observations, the above problems do not occur. Therefore, this paper proposes a new robust alternative method for deformation analysis of FCN (REDOD). The basis of this method is the assumption that vector d is a function of differences in independent observations \((\mathbf{d}=f(\mathbf{l}_{(2)}^\mathrm{obs} -\mathbf{l}_{(1)}^\mathrm{obs} ))\), i.e.,

$$\begin{aligned} \left( {\mathbf{l}_{(2)}^\mathrm{obs} -\mathbf{l}_{(1)}^\mathrm{obs}}\right) \, ^{\underrightarrow{\hbox {robust estimation}}}\,{\hat{\mathbf{d}}}^{R} \end{aligned}$$

If observations in both epochs are additionally distorted by constant errors, such an approach can eliminate the effect of these errors from deformation analysis results. As a result of estimation, the vector of differences in observations is assigned only such a vector of residuals which enables determination of vector \({\hat{\mathbf{d}}}^{R}\), but “does not bend” observations to fulfill the conditions for the closing of geometric figures. An unjustified increase in the value of the factor \(\hat{{\sigma }}_0^2\) (14) and distortions of vector \({\hat{\mathbf{d}}}^{R}\) are avoided this way. Moreover, the vector \({\hat{\mathbf{d}}}^{R}\) and the factor \(\hat{{\sigma }}_0^2 \) are estimated in the same computational process and, as a consequence, have the same stochastic properties.

The algorithm of the proposed REDOD method consists of the following two steps.

3.1 Step 1: robust estimation of the displacement vector

It can be seen that if the same geometric network elements have been measured at both analyzed epochs, then \(\mathbf{A}_{(1)} =\mathbf{A}_{(2)} =\mathbf{A}\), e.g., for continuous control measurements. In this case, it is possible to construct a certain alternative and very natural deformation model, in which the vector of observation differences will be the vector of “observations” and the sought displacement vector will be the vector of parameters. If interepoch observations, which are described by the model (2), are subtracted from each other, we will obtain for two epochs \(\left( {\mathbf{l}_{(2)}^\mathrm{obs} -\mathbf{l}^{0}} \right) +\mathbf{v}_{(2)} -\left( {\mathbf{l}_{(1)}^\mathrm{obs} -\mathbf{l}^{0}} \right) -\mathbf{v}_{(1)} ={\mathbf{A}{\varvec{\updelta }}}_{\mathbf{x}_{(2)}} -{\mathbf{A}{\varvec{\updelta }}}_{\mathbf{x}_{(1)}} \) and then

$$\begin{aligned} \left( {\mathbf{l}_{(2)}^\mathrm{obs} -\mathbf{l}_{(1)}^\mathrm{obs}} \right) \!+\!\mathbf{v}_{\varvec{\Delta }} \!=\!\mathbf{Ad},\,\left( {\mathbf{l}_{(2)}^\mathrm{obs} \!-\!\mathbf{l}_{(1)}^\mathrm{obs}} \right) \sim {N}\left( {\mathbf{Ad},\mathbf{C}_{\varvec{\Delta }}} \right) \end{aligned}$$
(17)

because \({\varvec{\updelta }}_{\mathbf{x}_{(2)}} -{\varvec{\updelta }}_{\mathbf{x}_{(1)}} =\mathbf{d},\,\mathbf{v}_{\varvec{\Delta }} =\mathbf{v}_{(2)} -\mathbf{v}_{(1)},\,\mathbf{C}_\Delta \in \mathfrak {R}^{n\times n}\) is the covariance matrix of differences in observations (Lazzarini et al. 1977). Because the vector of displacement components of FCN points may have a contaminated normal distribution with a mean of zero and accepted standard deviation, the solution to the system of Eq. (17) should be a robust estimator of vector d. In the REDOD method, as in the IWST method, the \(\hbox {L}_{1}\)-norm estimator of vector d will be used for this purpose. Moreover, because errors of geodetic observations are of random nature (according to normal distribution), the robust estimator of vector d should simultaneously fulfill the optimization condition of the LS method for the vector of residuals of differences in observations. Consequently, the following optimization problem for the system of Eq. (17) is obtained

(18)

where \(\mathbf{v}_\Delta \in \mathfrak {R}^{n\times 1}\) is the vector of residuals of differences in observations, \(\mathbf{Q}_\Delta =(\mathbf{Q}_{(1)} +\mathbf{Q}_{(2)} )\in \mathfrak {R}^{n\times n}\) is the cofactor matrix of differences in observations and \(\mathbf{P}_\Delta =\mathbf{Q}_\Delta ^{-1}\in \mathfrak {R}^{n\times n}\) is the weight matrix of differences in observations.

To find the minimum of the first objective function \(\phi \left( \mathbf{d} \right) \), the classical technique based on the differential calculus can be applied. Determining the gradient of the objective function \(\phi \left( \mathbf{d} \right) \) and then using the necessary condition for the existence of a minimum of this function \(\left( {\partial \phi \left( \mathbf{d} \right) /\partial \mathbf{d}} \right) ^\mathrm{T}=\mathbf{0}\), the following system of normal equations is obtained

$$\begin{aligned} \mathbf{N}_\Delta \mathbf{d}-{\varvec{{\upomega }}}_\Delta =\mathbf{0}, \end{aligned}$$
(19)

where \(\mathbf{N}_\Delta =\mathbf{A}^\mathrm{T}\mathbf{P}_\Delta \mathbf{A},\,{\varvec{{\upomega }}}_\Delta =\mathbf{A}^\mathrm{T}\mathbf{P}_\Delta \left( {\mathbf{l}_{(2)}^\mathrm{obs} -\mathbf{l}_{(1)}^\mathrm{obs}} \right) \). The objective function \(\phi \left( \mathbf{d} \right) \) is a convex function. Therefore, if the necessary condition is fulfilled, then the sufficient condition for a minimum of this function is fulfilled as well. The obtained system of normal Eq. (19) is, in fact, the set of permissible solutions (constraints) for vector \(\mathbf{d}\), minimizing the objective function \(\phi \left( \mathbf{d} \right) \).

Continuing the solution of the optimization problem (18), we find the minimum of the second objective function \(\psi (\mathbf{d})\), taking into account the constraints (19). The function \(\psi (\mathbf{d})\) is not repeatedly differentiable and its minimum cannot be found using the classical technique, based on differential calculus, as minimization of this objective function requires the use of complex linear programming procedures (e.g., simplex method). For this reason, in this paper, the function \(\psi (\mathbf{d})\) was replaced with an alternative, repeatedly differentiable function with the form (Kadaj 1988)

$$\begin{aligned} \psi _{\mathrm{A}} (\mathbf{d})=\sum _{i=1}^u {\sqrt{d_i^2 +c^{2}}}, \end{aligned}$$
(20)

where \(c\) is the precision for computations, e.g., 0.0001 m. With the assumption that \(c\) is a numerically insignificant value, the component of the function \(\psi _{\mathrm{A}} (\mathbf{d})\) is equivalent to the component of the function \(\psi (\mathbf{d})\) and then both functions are equivalent in terms of optimization properties. Objective function \(\psi _{\mathrm{A}} (\mathbf{d})\) is now repeatedly differentiable and its minimum can be found using the classical technique, based on differential calculus. According to the optimization theory, the method of Lagrangian multipliers can be applied to find the minimum of the objective function \(\psi _{\mathrm{A}} (\mathbf{d})\) with the constraints (19). According to this method, the objective function \(\psi _{\mathrm{A}} (\mathbf{d})\) with the constraints (19) is replaced by a secondary objective function without constraints (Lagrangian function)

$$\begin{aligned} L(\mathbf{d})=\sum _{i=1}^u {\sqrt{d_i^2 +c^{2}}} -{\varvec{\upkappa }}^\mathrm{T}\left( {\mathbf{N}_\Delta \mathbf{d}-{\varvec{{\upomega }}}_\Delta } \right) , \end{aligned}$$
(21)

where \({\varvec{\upkappa }} \in \mathfrak {R}^{u\times 1}\) is the vector of Lagrangian multipliers. The necessary conditions for the existence of a minimum of the objective function \(\psi _{\mathrm{A}} (\mathbf{d})\) with the constraints (19) then have the following form (Kuhn–Tucker necessary conditions)

$$\begin{aligned}&\left( {\partial L\left( \mathbf{d} \right) /\partial \mathbf{d}} \right) ^\mathrm{T}=\mathbf{0} \end{aligned}$$
(22)
$$\begin{aligned}&{\varvec{\upkappa }}^\mathrm{T}\left( {\mathbf{N}_\Delta \mathbf{d}-{\varvec{{\upomega }}}_\Delta } \right) =\mathbf{0} \end{aligned}$$
(23)

The objective function \(\psi _{\mathrm{A}} (\mathbf{d})\) is a convex function. Therefore, if the above necessary conditions are fulfilled, the sufficient conditions for a minimum of this function are fulfilled as well. Determining the gradient of the function \(L(\mathbf{d})\) and then using the necessary condition (22), the following equation is obtained

$$\begin{aligned} \mathbf{W}\hat{\mathbf{d}}-\mathbf{N}_\Delta {\hat{{\varvec{\upkappa }}}}=\mathbf{0} \end{aligned}$$
(24)

where

$$\begin{aligned} \mathbf{W}=\hbox {diag}\left( {\ldots ,\,1\big /\sqrt{\hat{{d}}_i^2 +c^{2}},\, \ldots } \right) \end{aligned}$$
(25)

and \(\mathbf{N}_{\Delta }^\mathrm{T} =\mathbf{N}_\Delta ,\,\mathbf{W}^\mathrm{T}=\mathbf{W}\), because \(\mathbf{N}_\Delta \) and \(\mathbf{W}\) are symmetrical matrices. The second necessary condition (23) can be replaced by the condition (19), (for \({\hat{{\varvec{\upkappa }}}}\ne \mathbf{0})\). Finally, after determination of vector \({\hat{\mathbf{d}}}\) from Eq. (24), the necessary condition (23) will assume the following form (system of normal equations of correlates)

$$\begin{aligned} {\mathbf{M}\hat{{\varvec{\upkappa }}}}-{\varvec{{\upomega }}}_\Delta =\mathbf{0}, \end{aligned}$$
(26)

where \(\mathbf{M}=\mathbf{N}_\Delta \mathbf{W}^{-1}\mathbf{N}_\Delta \). If vector \({\hat{{\varvec{\upkappa }}}}\) solves the system of Eq. (26), vector \({\hat{\mathbf{d}}}\) minimizes the objective function \(\psi _{\mathrm{A}} (\mathbf{d})\) with the constraints (19). Therefore, after determination of vector \({\hat{{\varvec{\upkappa }}}}\) from Eq. (26) and its substitution into Eq. (24), we finally obtain

$$\begin{aligned} \mathbf{W}\hat{\mathbf{d}}-\mathbf{N}_\Delta \mathbf{M}^{-}{\varvec{{\upomega }}}_\Delta =\mathbf{0}, \end{aligned}$$
(27)

where \(\mathbf{M}^{-}\) is a g-inverse, such that \(\mathbf{MM}^{-}\mathbf{M}=\mathbf{M}\). In the considered case of an FCN, A is the matrix of the columnarly incomplete rank (4), which is why the matrix M is singular and there is no ordinary (classical) inverse \(\mathbf{M}^{-1}\). To determine the g-inverse \(\mathbf{M}^{-}\), according to g-inverse theory, the matrix \(\mathbf{M}\in \mathfrak {R}^{u\times u}\) can be reduced to block form

$$\begin{aligned} \mathbf{M}=\left[ {{\begin{array}{ll} {\mathbf{M}_{11}} &{}\quad {\mathbf{M}_{12}} \\ {\mathbf{M}_{21}} &{}\quad {\mathbf{M}_{22}} \end{array}}} \right] \rightarrow \mathbf{M}^{-}=\left[ {{\begin{array}{ll} {\mathbf{M}_{11}^{-1}} &{} \quad \mathbf{0} \\ \mathbf{0}&{}\quad \mathbf{0} \end{array}}} \right] , \end{aligned}$$
(28)

where \(\mathbf{M}_{11} \in \mathfrak {R}^{r\times r}\) is a non-singular matrix with rank \(r\). The g-inverse of the matrix M is then the submatrix \(\mathbf{M}_{11}^{-1} \), complemented with zeros to the dimension \(\mathfrak {R}^{u\times u}\) (Rao 1973; Prószyński 1981). Of course, there are also other methods for determination of the g-inverse of this matrix. It can, therefore, be seen that we have here the problem of a datum defect for the geodetic network, which has not yet been solved. To reduce the matrix M to the form (28), it is sufficient to represent the matrix A in block form (5). Therefore, in this case, the problem of a datum defect for the geodetic network can be solved analogously as in the classical method, only in this case it is done simultaneously for both epochs. Moreover, let us recall that the same optimization conditions were also proposed in both methods. It can, therefore, be seen that both methods differ only in the functional models, i.e., (9) and (18), while the other theoretical assumptions are common. Returning to the computations, after substitution of the block matrix A (5) into the matrix M, we obtain

$$\begin{aligned} \mathbf{M}^{-}=\left[ {{\begin{array}{ll} {(\mathbf{Q}_{11} \mathbf{W}_1^{-1} \mathbf{Q}_{11} +\mathbf{Q}_{12} \mathbf{W}_2^{-1} \mathbf{Q}_{12}^\mathrm{T} )^{-1}}&{}\quad \mathbf{0} \\ \mathbf{0}&{}\quad \mathbf{0} \end{array}}}\right] , \end{aligned}$$
(29)

where \(\mathbf{Q}_{11} =\mathbf{A}_1^\mathrm{T} \mathbf{P}_\Delta \mathbf{A}_{1},\,\mathbf{Q}_{12} =\mathbf{A}_1^\mathrm{T} \mathbf{P}_\Delta \mathbf{A}_2 \) and \(\mathbf{W}=\hbox {diag}(\mathbf{W}_1 ,\mathbf{W}_2)\). After substitution of the matrix (29) into the system of Eq. (27), we finally obtain

$$\begin{aligned} \mathbf{W}\hat{\mathbf{d}}-\mathbf{D}^\mathrm{T}\left( {\mathbf{DW}^{-1}\mathbf{D}^\mathrm{T}} \right) ^{-1}\mathbf{A}_1^T \mathbf{P}_\Delta \left( {\mathbf{l}_{(2)}^\mathrm{obs} -\mathbf{l}_{(1)}^\mathrm{obs}}\right) =\mathbf{0} \end{aligned}$$
(30)

where \(\mathbf{D}= [{ \begin{array}{l@{\quad }l} {\mathbf{A}_1^\mathrm{T} \mathbf{P}_\Delta \mathbf{A}_1}&{\mathbf{A}_1^\mathrm{T} \mathbf{P}_\Delta \mathbf{A}_2} \end{array}}]\). We compute on the basis of the system of Eq. (30) the robust estimator of vector d and its cofactor matrix. The computations are carried out in an iterative cycle with the following formula (approximate Newtonian process)

$$\begin{aligned} \left. {\begin{array}{c} {\hat{\mathbf{d}}}^{(k)}=\mathbf{R}^{(k)}\left( {\mathbf{l}_{(2)}^\mathrm{obs} -\mathbf{l}_{(1)}^\mathrm{obs}} \right) \\ \mathbf{Q}_{{\hat{\mathbf{d}}}}^{(k)} =\mathbf{R}^{(k)}\left( {\mathbf{Q}_{(1)} +\mathbf{Q}_{(2)}} \right) \left( {\mathbf{R}^{(k)}} \right) ^\mathrm{T} \\ \mathbf{W}^{(k+1)}=\hbox {diag}\left( {\ldots ,\, w_i^{(k+1)} ,\,\ldots } \right) \end{array}} \right\} _{k=1,2,\ldots } \end{aligned}$$
(31)

where \(\mathbf{R}^{(k)}=\left( {\mathbf{W}^{(k)}}\right) ^{-1}\mathbf{D}^\mathrm{T}\left( {\mathbf{D}\left( {\mathbf{W}^{(k)}}\right) ^{-1}\mathbf{D}^\mathrm{T}} \right) ^{-1}\mathbf{A}_1^\mathrm{T} \mathbf{P}_\Delta \) is the matrix of transformation of the vector of differences in observations into the displacement vector, \(\mathbf{W}^{(k=1)}=\mathbf{I}\) and

$$\begin{aligned} w_i^{(k+1)}=1\big /\sqrt{\left( {\hat{{d}}_i^{(k)}} \right) ^{2}+c^{2}}\, \cong \,1\big /\left( {\left| {\hat{{d}}_i^{(k)}}\right| +c}\right) \end{aligned}$$
(32)

It should be stressed that there still remains problem of outlier detection (observations), which has not yet been solved. In this case, this problem can be solved, e.g., on the basis of Baarda’s data snooping or Pope’s test, only in this case it is done on the basis of the vector of residuals of differences in observations from the first iterative step (31). Unfortunately, in this case, robust estimation “does not work” for detecting outliers (observations).

The iterative process (31) is finished when all components of the difference \(\left| {{\hat{\mathbf{d}}}^{(k+1)}-{\hat{\mathbf{d}}}^{(k)}} \right| \) are lower than the adopted precision for computations \(c\) (convergence). The final robust estimator of the vector of displacement components for all analyzed points in an FCN, i.e., vector \({\hat{\mathbf{d}}}^{R}\) and the final cofactor matrix of this vector, i.e., matrix \(\mathbf{Q}_{{\hat{\mathbf{d}}}^{R}} \), is obtained from the last iterative step.

3.2 Step 2: assessment of the significance of the displacement vector

Assessment of the significance of the vectors \({\hat{\mathbf{d}}}_i^R \) can of course be performed analogously as in the classical method, i.e., on the basis of the approximate global (13) and local (15) or (16) \(F\) tests. For the stochastic model of the REDOD method (18), the variance factor estimator has the following form

$$\begin{aligned} \hat{{\sigma }}_0^2 ={\hat{\mathbf{v}}}_\Delta ^\mathrm{T} \mathbf{P}_\Delta {\hat{\mathbf{v}}}_\Delta /(n-u+de) \end{aligned}$$
(33)

4 Comparison of the classical IWST method with the proposed REDOD method

As previously mentioned, classical deformation analysis of FCN is most often performed by robust methods. The best known methods from this group include the IWST method. This method was used, among others, for deformation analysis of the Tevatron atomic particle accelerator complex at the Fermilab laboratory in the US and for deformation analysis of the world’s largest strip copper mine in Chile. The IWST method has also been implemented in the automated ALERT monitoring system developed by the Canadian Centre for Geodetic Engineering and in the universal GeoLab geodetic computation software. Section 2 describes the algorithm of classical robust methods. Particular attention was paid to the IWST method. In Sect. 3, a new robust deformation analysis method was proposed based on classical robust methods. This method was conventionally called REDOD.

The theoretical principles of the proposed REDOD method are very similar to the theoretical principles of the classical IWST method. Both methods have in common, among others, the definition of the datum for the geodetic network, i.e., the zero-variance computational base (5), the optimization condition for the displacement vector (the definition of the datum for the displacement vector), i.e., the minimal \(\hbox {L}_{1}\)-norm (9, 18), or the probabilistic model for the observations, i.e., normal distribution (2, 17).

The theoretical principles of both methods differ only in the functional deformation models, i.e., (9) and (18). Generally, the functional models (9) and (18) assume that the displacement vector is a function of differences in adjusted coordinates and differences in unadjusted observations, respectively. It is worth noting once more that the functional model (18) has some advantages which the functional model (9) does not have, namely, the functional model (18) accepts measurement errors constant in both epochs, which can be of any magnitude and may be different for individual geometric elements of the network. Moreover, unlike model (9), model (18) ensures stochastic consistency between the displacement vector estimator and the variance factor estimator. This results from the fact that both estimators are estimated here in the same computational process.

As a consequence of using different functional models, the aithms of the IWST and REDOD methods differ significantly. In Sects. 2 and 3, these algorithms were derived and described in detail. Only their basic aspects are shown below, in the form of a block diagram (Fig. 1).

Fig. 1
figure 1

Block diagram of the algorithms of the IWST and REDOD methods

5 Numerical tests

Numerical tests were performed on simulated FCN. The tests compared deformation analysis results obtained from the IWST and REDOD methods.

5.1 Example 1: deformation analysis of the leveling FCN

This test was performed on the example of a leveling network of four points A, B, C, D (Fig. 2). Computations were carried out for two variants of simulated observations.

Fig. 2
figure 2

Leveling FCN

First, the vector of theoretical heights of controlled points in epoch 1 \(\mathbf{x}_{(1)}= [0.000,~0.000,~0.000,~0.000]^\mathrm{T}\hbox { m}\), the vector of theoretical displacements of controlled points \(\mathbf{d} = [0,~10,~0,~0]^\mathrm{T}\hbox { mm}\) and the vector of theoretical heights of controlled points in epoch 2 \(\mathbf{x}_{(2)}=[0.000,~0.010,~0.000,~0.000]^\mathrm{T}\hbox { m}\) were assumed. Theoretical height differences in both epochs were computed on the basis of the vectors \(\mathbf{x}_{(1)} \) and \(\mathbf{x}_{(2)}\).

In variant 1, the theoretical height differences were distorted with random errors according to the normal distribution and the significance level \(\alpha = 0.05\). The same standard deviations of simulated errors of height difference measurement \(\sigma _\mathrm{h} = 4~\hbox {mm}\) were assumed in both epochs.

In variant 2, the theoretical height differences were distorted with random errors from variant 1 and additionally with constant errors \(\varepsilon _\mathrm{h} = +2\,\hbox {mm}\).

The variants of simulated observations are presented in Table 1. Carrying out in succession the computations (10), (14), (13) and (15), complete results were obtained using IWST method. Carrying out in succession the computations (31), (33), (13) and (15) complete results were obtained using REDOD method. In both methods, constant heights \(z_D \) were assumed to eliminate the network’s datum defect [Eq. (3) in the classical method and Eq. (27) in the proposed method], i.e., the submatrix \(\mathbf{A}_2 \) consisted of the de = 1 last column in the matrix A, as was proposed in the theoretical part of this paper (5). Standard deviations of simulated errors of height difference measurement were assumed as a priori mean errors of height difference measurement. The values of a priori mean errors were the basis for determining the matrices \(\mathbf{P}_{(1)},\,\mathbf{P}_{(2)} \) (IWST) and \(\mathbf{P}_\Delta \) (REDOD). Outlier detection was carried out using Baarda’s method. The critical values were \(\chi _{0.05}^2 (3)/3=\hbox {2.61}\) and \(t_{\alpha _0 /2} =2.51\,(\beta _0 =0.20)\) for the global and local test, respectively. In variant 1, in the IWST method, the global test passed. In variant 2, in the IWST method, the global test failed in epoch 2 \((\hat{{\sigma }}_{0(2)}^2 =2.89>2.61)\), but the local test passed for all standardized residuals. In the REDOD method, the global test passed in both variants. In the IWST method, the test on the variance ratio passed in both variants [critical value \(F(0.05,3,3)=9.28]\).

Table 1 Simulated observations in epoch 1 and epoch 2

The results of the deformation analysis are presented in Table 2.

Table 2 The results of the deformation analysis of the leveling FCN

In variant 1, both methods give the same, correct results. In both methods, point B is considered unstable. In variant 2, in the IWST method, the value of vector \({\hat{\mathbf{d}}}^{R}\) is the same as in variant 1. Therefore, constant errors have no effect on vector \({\hat{\mathbf{d}}}^{R}\). However, the results of the assessment of significance are different than in variant 1. This is caused by the effect of constant errors. This effect was not eliminated from the values of the vectors \({\hat{\mathbf{v}}}_{(1)},\,{\hat{\mathbf{v}}}_{(2)}\). This led to an increase in the value of \(\hat{{\sigma }}_0^2 \) (14), underestimation of the values of \(T\) (13), \(T_i\) (15) and, consequently, erroneous assessment of the significance of vector \({\hat{\mathbf{d}}}_B^R \). In the REDOD method, the effect of constant errors is eliminated. Both the value of vector \({\hat{\mathbf{d}}}^{R}\) and the results of the assessment of significance are the same as in variant 1.

5.2 Example 2: deformation analysis of the horizontal FCN

This test was performed on the example of a horizontal network of six points A, B, C, D, E, F (Fig. 3). Computations were carried out for four variants of simulated observations.

Fig. 3
figure 3

Horizontal FCN

First, the vector of theoretical coordinates of controlled points in epoch 1 \(\mathbf{x}_{(1)}\)= [350.000, 200.000, 300.000, 300.000, 200.000, 300.000, 150.000, 200.000, 200.000, 100.000, 300.000, 100.000]\(^\mathrm{T}\) m, the vector of theoretical displacements of controlled points d = [5, 5, 0, 0, 0, 10, 0, 0, 0, 0, 0, 0]\(^\mathrm{T}\hbox { mm}\) and the vector of theoretical coordinates of controlled points in epoch 2 \(\mathbf{x}_{(2)}\) = [350.005, 200.005, 300.000, 300.000, 200.000, 300.010, 150.000, 200.000, 200.000, 100.000, 300.000, 100.000]\(^\mathrm{T}\hbox { m}\) were assumed.

In variant 1, observations were distorted with random errors. Theoretical observations were computed on the basis of the vectors \(\mathbf{x}_{(1)},\,\mathbf{x}_{(2)} \). The theoretical observations were then distorted with random errors according to the normal distribution and the significance level \(\alpha = 0.05\). The same standard deviations of simulated errors of angle measurement \(\sigma _{\beta } =10^{\mathrm{cc}}\) and distance measurement \(\sigma _{s} =~3~\hbox {mm}\) were assumed in both epochs.

In variant 2, observations were also distorted with random errors. Theoretical observations were computed on the basis of the vectors \(\mathbf{x}_{(1)},\,\mathbf{x}_{(2)} \). The theoretical observations were then distorted with random errors according to the normal distribution and the significance level \(\alpha = 0.05\). However, different standard deviations of simulated errors of angle measurement were assumed in both epochs in this case. In epoch 1, the standard deviation of angle measurement errors at central points A, B, C \(\sigma _{\beta } =~15^{\mathrm{cc}}\), at central points D, E, F \(\sigma _{\beta } =~5^{\mathrm{cc}}\) and the standard deviation of distance measurement errors \(\sigma _{s}= 3~\hbox {mm}\) were assumed. In epoch 2, the standard deviation of angle measurement errors at central points A, B, C \(\sigma _{\beta }= 5^{\mathrm{cc}}\), at central points D, E, F \(\sigma _{\beta }= 15^{\mathrm{cc}}\) and the standard deviation of distance measurement errors \(\sigma _{s} = 3~\hbox {mm}\) were assumed.

In variant 3, observations were distorted with random errors and constant errors. The vector \(\varepsilon _\mathbf{x}\) = [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3]\(^\mathrm{T}\hbox { mm}\) was added to the vectors \(\mathbf{x}_{(1)},\,\mathbf{x}_{(2)} \) and the vectors of eccentric coordinates of the target \(\mathbf{x}_{(1)}^\varepsilon ,\,\mathbf{x}_{(2)}^\varepsilon \) were obtained. The eccentric observations were then computed on the basis of the theoretical coordinates of the instrument \(\mathbf{x}_{(1)},\,\mathbf{x}_{(2)} \) and the eccentric coordinates of the target \(\mathbf{x}_{(1)}^\varepsilon ,\,\mathbf{x}_{(2)}^\varepsilon \). In other words, observations distorted by constant errors in the form of eccentricity of the target in relation to the instrument with the value of \(\varepsilon _\mathbf{x} = +3~\hbox {mm}\) along the axes \(X\) and \(Y\) were obtained this way. These observations were then additionally distorted with random errors from variant 1.

In variant 4, observations were distorted with constant errors from variant 3 and random errors from variant 2.

The variants of simulated observations are presented in Table 3.

Table 3 Simulated observations in epoch 1 and epoch 2

Complete results were obtained in the IWST method by carrying out the computations (10), (14), (13) and (16) in succession. Complete results were obtained in the REDOD method by carrying out the computations (31), (33), (13) and (16) in succession. In both methods, constant coordinates \(y_\mathrm{E},\,x_\mathrm{F},\,y_\mathrm{F} \) were assumed to eliminate the network’s datum defect (Eq. (3) in the classical method and Eq. (27) in the proposed method), i.e., the submatrix \(\mathbf{A}_2 \) consisted of the de = 3 last columns in the matrix A, as was proposed in the theoretical part of this paper (5). The theoretical coordinates from epoch 1 were assumed as approximate coordinates. Standard deviations of simulated errors of angle and distance measurement were assumed as a priori mean errors of angle and distance measurement. The values of a priori mean errors were the basis for determining the matrices \(\mathbf{P}_{(1)},\,\mathbf{P}_{(2)} \) (IWST) and \(\mathbf{P}_\Delta \) (REDOD). Outlier detection was carried out by Baarda’s method. The critical values were \(\chi _{0.05}^2 (17)/17=\hbox {1.62}\) and \(t_{\alpha _0 /2} =3.63\,(\beta _0 =0.20)\) for the global and local test, respectively. The global test passed in variants 1 and 2 in the IWST method, but the global test failed in both epochs in variant 3 in the IWST method \((\hat{{\sigma }}_{0(1)}^2 =2.51>1.62,\,\hat{{\sigma }}_{0(2)}^2 =2.10>1.62)\), although the local test passed for all standardized residuals. In variant 4, in the IWST method, the global test failed in both epochs \((\hat{{\sigma }}_{0(1)}^2 =2.98>1.62,\,\hat{{\sigma }}_{0(2)}^2 =3.42>1.62)\) and the local test failed in both epochs. The largest standardized residuals of epochs 1 and 2 were \(\hat{{v}}_\mathrm{DEF} /\sigma _{\hat{{v}}_\mathrm{DEF}} =4.51>3.63\) and \(\hat{{v}}_\mathrm{ABC} /\sigma _{\hat{{v}}_\mathrm{ABC}} =5.36>3.63\), respectively. These observations were deleted. The critical values at that point were \(\chi _{0.05}^2 (16)/16=\hbox {1.64}\) and \(t_{\alpha _0 /2} =3.58\,(\beta _0 =0.20)\) for the global and local test, respectively. In the IWST method, the global test failed again in both epochs \((\hat{{\sigma }}_{0(1)}^2 =1.98>1.64,\,\hat{{\sigma }}_{0(2)}^2 =1.92>1.64)\), but the local test passed now for all standardized residuals. In the REDOD method, the global test passed in all variants. In the IWST method, the test on the variance ratio passed in all variants [critical value \(F(0.05,17,17)=2.27]\).

The results of deformation analysis are presented in Table 4.

Table 4 The results of deformation analysis of the horizontal FCN

In variant 1, both methods yield the same results (within the limits of accuracy of computations).

In variant 2, the values of vector \({\hat{\mathbf{d}}}^{R}\) and the parameters of confidence ellipses differ in both methods. This is caused by disproportionateness of the matrices \(\mathbf{P}_{(1)},\,\mathbf{P}_{(2)} \) (IWST) to the matrix \(\mathbf{P}_\Delta \) (REDOD). However, these differences are not significant. Moreover, the sum of absolute values of true errors for the components of vector \({\hat{\mathbf{d}}}^{R}\) is the same in both methods, \(\sum {\left| {\hat{{d}}_i^R -d_i} \right| } =8.2\hbox { mm}\).

In variant 3, in the IWST method, the value of vector \({\hat{\mathbf{d}}}^{R}\) is the same as in variant 1. Therefore, constant errors have no effect on vector \({\hat{\mathbf{d}}}^{R}\). However, the values of the parameters of confidence ellipses differ significantly from the values in variant 1. This is caused by the effect of constant errors. This effect was not eliminated from the values of the vectors \({\hat{\mathbf{v}}}_{(1)},\,{\hat{\mathbf{v}}}_{(2)} \). This led to an increase in the value of \(\hat{{\sigma }}_0^2 \) (14) and, consequently, to an increase in the values of the parameters of confidence ellipses (16). In the REDOD method, the effect of constant errors is eliminated. Both the values of vector \({\hat{\mathbf{d}}}^{R}\) and the values of the parameters of confidence ellipses are the same as in variant 1.

In variant 4, in the IWST method, the value of vector \({\hat{\mathbf{d}}}^{R}\) differs significantly from the value in variant 2. The sum of absolute values of true errors for the values of vector \({\hat{\mathbf{d}}}^{R}\) is significantly higher than in variant 2, \(\sum {\left| {\hat{{d}}_i^R -d_i} \right| } =26.9\hbox { mm}\). This is caused by the disproportion of matrix \(\mathbf{P}_{(1)} \) to matrix \(\mathbf{P}_{(2)} \). This in turn led to a disproportionate effect of constant errors on the values of the vectors \({\hat{\mathbf{x}}}_{(1)},\,{\hat{\mathbf{x}}}_{(2)} \) (1) and, consequently, prevented complete reduction of these errors in the iterative process (10). The values of the parameters of confidence ellipses also differ significantly from the values in variant 2. As in variant 3, this is caused by the effect of constant errors on the value of \(\hat{{\sigma }}_0^2 \). In the REDOD method, the effect of constant errors is eliminated. Both the values of vector \({\hat{\mathbf{d}}}^{R}\) and the values of the parameters of confidence ellipses are the same as in variant 2.

Figure 4 presents graphical assessment of the significance of vector \({\hat{\mathbf{d}}}^{R}\) (16).

Fig. 4
figure 4

The results of assessment of the significance of displacements in the case of the occurrence of random errors (variants 1, 2) and in the case of the occurrence of random and constant errors (variants 3, 4)

In the case of the occurrence of only random errors (Fig. 4, variants 1, 2), the results of assessment of significance are correct in both methods. The displacements of points A and C are considered significant in both variants.

In the case of the occurrence of random errors and constant errors (Fig. 4, variants 3, 4), the results of assessment of significance in the IWST method are not correct. Only the displacement of point C is considered significant in both variants. This is caused by the effect of constant errors. In the REDOD method, the results of assessment of significance are correct. The displacements of points A and C are considered significant in both variants.

6 Summary and conclusions

Deformation measurements have a repeatable nature. In other words, deformation measurements are performed at different moments of time, but often with the same equipment, methods, geometric conditions and in a similar environment in epochs 1 and 2. Moreover, the same instrument and target are most often set on given points in both epochs. It is, therefore, reasonable to assume that the results of deformation measurements can be distorted both by random errors and by some non-random errors, which are constant in both epochs, i.e., errors with the same value and the same sign in both epochs. There may be constant errors which cannot be eliminated from measurement results, e.g., incompletely reduced atmospheric refraction, eccentricity of the target in relation to the instrument (caused by faulty execution or assembly, non-vertical mounting of the centering sleeve in the instrument pillar or inclination of the instrument pillar with time). There may also be constant errors which can be eliminated from measurement results although, for example, economic and technical limitations of the measurement project prevent their elimination (Caspary 2000). For example, a pillar could have a GNSS mounted, as well as a target prism that RTS could observe (Fig. 5). In this case, the target prism and GNSS receiver cannot define the same physical point.

Fig. 5
figure 5

Integrated deformation measurements (RTS, target prism and GNSS receiver) for network of potential reference points (FCN)

For FCN, the results of deformation measurements are most often processed using the robust methods. Classical robust methods do not completely eliminate the effect of constant errors and, therefore, lead to reduced external reliability level of processing and, consequently, to reduced quality of deformation analysis results.

This paper proposes a new robust alternative method, which was called REDOD. The performed numerical tests showed that:

  1. 1.

    If the results of deformation measurements were additionally distorted by constant errors, the REDOD method completely eliminated their effect from deformation analysis results,

  2. 2.

    If the results of deformation measurements were distorted only by random errors, the REDOD method yielded very similar deformation analysis results as the classical IWST method.

The REDOD method was developed for FCN, which has the same observation structure in both epochs. That is, the same geometric elements must be measured in the control network in both epochs. For example, this condition is always met for automated, continuous control measurements. As previously mentioned, there is a high risk that the error of measurement of the same geometric element of the network in both epochs may contain a factor constant in both epochs and, therefore, the REDOD method is strongly recommended in this case.

It should be noted, however, that sometimes, in practice, control networks have a different observation structure in both epochs. For example, the classical technique in epoch 1 can be replaced with the GNSS technique in epoch 2. In this case, the REDOD method can also be used. The input data for the REDOD method will then be pseudo-observations of the same geometric elements of the network in both epochs (e.g., vectors). Such pseudo-observations can be computed based on preadjusted coordinates. However, it should be noted that for networks which have different observation structure in both epochs, there is a very low risk that error constant in both epochs will occur and, therefore, in this case, the REDOD method is not strongly recommended. As previously demonstrated, if there are no error constant in both epochs, the REDOD and IWST methods give much the same quality of deformation analysis results.

he obtained results encourage further research into the latest robust solutions for deformation estimation from observation differences. The method of sign-constrained robust least squares, which detects over 50 % of outliers (Xu 2005), and robust estimation by expectation maximization algorithm, which is more sensitive to outliers than classical methods (Koch 2013; Koch and Kargoll 2013), may be particularly useful. Moreover, new numerical tests should be performed for new approaches, based on many thousands of simulated data sets (e.g., Monte Carlo simulations). Not only measurement errors should be randomized in individual sets, but also the number and location of displaced points and the values and directions of these points’ displacements. The measure of accuracy of the estimated displacements could be empirical standard deviation and the measure of significance test efficiency could be the mean success rate (MSR) (Hekimoglu and Koch 1999).