next up previous [pdf]

Next: Multidimensional recursive filter preconditioning Up: One-dimensional synthetic examples Previous: One-dimensional synthetic examples

Regularization after binning: missing data interpolation

One of the factors affecting the convergence of iterative data regularization is clustering of data points in the output bins. When least-squares optimization assigns equal weight to each data point, it may apply inadequate effort to fit a cluster of data points with similar values in a particular output bin. To avoid this problem, we can replace the regularized optimization with a less accurate but more efficient two-step approach: data binning followed by missing data interpolation.

Missing data interpolation is a particular case of data regularization, where the input data are already given on a regular grid, and we need to reconstruct only the missing values in empty bins. The basic principle of missing data interpolation is formulated as follows (Claerbout, 1992):

A method for restoring missing data is to ensure that the restored data, after specified filtering, has minimum energy.
Mathematically, this principle can be expressed by the simple equation
\begin{displaymath}
\mathbf{D m \approx 0}\;,
\end{displaymath} (19)

where $\mathbf{m}$ is the model vector and $\mathbf {D}$ is the specified filter. Equation (19) is completely equivalent to equation (2). The approximate equality sign means that equation (19) is solved by minimizing the squared norm (the power) of its left side. Additionally, the known data values must be preserved in the optimization scheme. Introducing the mask operator $\mathbf{K}$, which is a diagonal matrix with zeros at the missing data locations and ones elsewhere, we can rewrite equation (19) in the extended form
\begin{displaymath}
\mathbf{D (I-K) m} \approx - \mathbf{D K m} = - \mathbf{D m}_k\;,
\end{displaymath} (20)

where $\mathbf{I}$ is the identity operator and $\mathbf{m}_k$ represents the known portion of the data. It is important to note that equation (20) corresponds to the regularized linear system
\begin{displaymath}
\left\{\begin{array}{l}
\mathbf{K m} = \mathbf{m}_k\;, \\
\epsilon \mathbf{D m} \approx \mathbf{0}
\end{array}\right.
\end{displaymath} (21)

in the limit of the scaling coefficient $\epsilon$ approaching zero. System (21) is equivalent to system (1-2) with the masking operator $\mathbf{K}$ playing the role of the forward interpolation operator $\mathbf{L}$. Setting $\epsilon$ to zero implies associating more weight on the first equation in (21) and using the second equation only to constrain the null space of the solution. Applying the general theory of data-space regularization from the previous section, we can immediately transform system (21) to the equation
\begin{displaymath}
\mathbf{K P p} \approx \mathbf{m}_k\;,
\end{displaymath} (22)

where $\mathbf {P}$ is a preconditioning operator and $\mathbf{p}$ is the preconditioning variable, connected with $\mathbf{m}$ by the simple relationship (8). According to equations (14) and (7) from the previous section, equations (22) and (20) have exactly the same solutions if condition (15) is satisfied. If $\mathbf {D}$ is represented by a discrete convolution, the natural choice for $\mathbf {P}$ is the corresponding deconvolution (inverse recursive filtering) operator:
\begin{displaymath}
\mathbf{P} = \mathbf{D}^{-1}\;.
\end{displaymath} (23)

We illustrate the missing data problem with a simple 1-D synthetic data test taken from Claerbout (1999). Figure 8 shows the interpolation results of the unpreconditioned technique with two different filters $\mathbf {D}$. For comparison with the preconditioned scheme, we changed the boundary convolution conditions from internal to truncated transient convolution. As in the previous example, the system was solved with a conjugate-gradient iterative optimization.

mall
mall
Figure 8.
Unpreconditioned interpolation with two different regularization filters. Left plot: the top shows the input data; the middle, the result of interpolation; the bottom, the filter impulse response. The right plot shows the convergence process for the first four iterations.
[pdf] [png] [scons]

As depicted on the right side of the figures, the interpolation process starts with a ``complicated'' model and slowly extends it until the final result is achieved.

Preconditioned interpolation behaves differently (Figure 9). At the early iterations, the model is simple. As the iteration proceeds, new details are added into the model. After a surprisingly small number of iterations, the output closely resembles the final output. The final output of interpolation with preconditioning by recursive deconvolution is exactly the same as that of the original method.

sall
sall
Figure 9.
Interpolation with preconditioning. Left plot: the top shows the input data; the middle, the result of interpolation; the bottom, the filter impulse response. The right plot shows the convergence process for the first four iterations.
[pdf] [png] [scons]

The next section extends the idea of preconditioning by inverse recursive filtering to multiple dimensions.


next up previous [pdf]

Next: Multidimensional recursive filter preconditioning Up: One-dimensional synthetic examples Previous: One-dimensional synthetic examples

2013-03-03