Professional Documents
Culture Documents
In practical non-linear problems, the data and the model parameter vectors are related via a
non-linear response function f, which tells us how to calculate the synthetic data from the
given model. This inter-relationship is expressed as
d=f(m)+e (2.1)
so that
(m)=eTe=(df(m))T(df(m)), (2.2)
where e is the N1 column vector of residuals [Lines and Treitel, 1984]. The overall goal of
the inversion is then to minimize the deviation between the measured and calculated data
(m), what may be described as the objective function.
To derive an iterative inversion scheme, the response function f is linearized about a starting
model (initial guess) mo by expanding it into a Taylor's series approximation and ignoring the
higher terms as
f(m)f(mo)+Jm with m=mom, (2.3)
where m is the model update (or perturbation) vector and J is the Jacobian matrix.
The NM Jacobian matrix J is a by-product of linearizing a non-linear problem. Its elements
are the partial derivatives of the calculated data with respect to the model parameters (i.e.
layer resistivities and thicknesses)
fi(m)
Jij= , i= 1, 2, 3, ..., N ; j= 1, 2, 3, ..., M.
mj m= mo
Because, each column indicates the operating RMT frequencies or TEM time windows in
which an individual parameter can be predominantly resolved, this matrix is called parameter
sensitivity matrix. The full space Jacobian can be written as
f1/m1 f1/m2 . . . f1/mM
f2/m1 f2/m2 . . . f2/mM
. . . ., (2.4)
J= . . .
. . .
. . .
. . .
fN/m1 fN/m2 . . . fN/mM
N M
The calculation of the Jacobian at each iteration is usually the most time-consuming part in
the inversion. Statistically, sensitivity values show whether the layer parameters are seen
individually in the measured data at all or, if not, they qualitatively display which a parameter
correlation is resolved by the data instead. Therefore, a model parameter is only well-resolved
by the measurements if the data points depend strongly on this parameter, i.e. if its sensitivity
has large entries in the corresponding column. On the other hand, the normalized Jacobian
values or relative sensitivities give an idea of each data point's sensitivity to a change of an
individual parameter in the solution [Jupp and Vozoff, 1975; Lines and Treitel, 1984].
If we define the vector y=df(mo) as the discrepancy vector, then the Equation (2.2) can be
written in the form
(m)=eTe=(yJm)T(yJm), (2.5)
16
Chapter 2 INVERSION OF ELECTROMAGNETIC DATA
In the present work, all inversion schemes attempt at minimizing the familiar root mean-
square (RMS) defined by
2 .
RMS= (2.9)
N
17
Chapter 2 INVERSION OF ELECTROMAGNETIC DATA
where . denotes the usual L2 (Euclidean) norm, mo is a finite quantity, and is the
MarquardtLevenberg damping factor [Marquardt, 1963; Levenberg, 1944], i.e. a varying
very small positive number allowed to numerically damp the model update at each iteration.
The minimization of the objective function leads to so-called damped least-squares solution
m=(JTJ+ 2I)1JTy (2.12)
or generally
mk+1=mk+(JTJ+ 2I)1JTy, (2.13)
where I is the MM identity matrix.
Incorporating the data weighting into the minimization problem (2.13) [Hrdt et al., 1992b]
changes the solution to
m=(JTW2J+ 2I)1JTW2y, (2.14)
where W is a NN waiting matrix, its diagonal elements are the reciprocal values of si.
There are several ways to evaluate the performance of the MarquardtLevenberg inversion
and reliability of results. The most favorable statistical approaches are either the singular-
value decomposition (SVD) analysis or the resolution matrix analysis. Typically, these
inversion statistics are only applicable for the best-fitted model, where the linearity
assumption is valid [Lines and Treitel, 1984].
Singular-value decomposition
A way of obtaining an evaluation of the parameter resolution is the additional use of the so-
called singular-value decomposition (SVD) [Jupp and Vozoff, 1975; Jackson, 1972]. The
weighted Jacobian Jw=WJ is substituted by the product of two orthogonal matrices V and U,
and one MM matrix S, its diagonal elements contain the positive roots of the eigenvalues of
Jw, Si.
Jw=WJ=USVT, (2.15)
where U is a NM matrix which contains the data space eigenvectors of JwTJw in its columns,
V is a MM matrix which contains the parameter space square-roots of the eigenvalues of
JwTJw in its columns (i.e. singular values of Jw). All matrices satisfy UTU=VTV=VVT=I,
18
Chapter 2 INVERSION OF ELECTROMAGNETIC DATA
where I is the NN identity matrix. Applying this matrix decomposition to the Equation
(2.14), we obtain
m=Vdiag(1/Si)diag(Si2/(Si2+ 2))UTWy. (2.16)
The Si2/(Si2+2) are the by-products of the SVD analysis, which are called the damping factors
of the inverted layer parameters. In the present work, an exponent of 4 was used instead of 2.
In this case, the damping factors act more like thresholds at each iteration. They are generally
measures of 'importances' of the corresponding layer parameters, and vary between 0 and 1 to
represent the minimal and maximal parameter resolution respectively.
Resolution matrix
For simplicity, we rewrite the solved inverse problem (2.14) in the general form of
mEst=Ggd, (2.17)
where Gg denotes the generalized inverse, which we can use to examine the resolution of the
estimated model mE s t , and therefore
mEst=GgGm=Rm, (2.18)
where R=G-gG is the parameter resolution matrix. It is a matrix which, when multiplied onto
the model vector m, determines the linear combination of the layer parameters that could be
resolved. Its diagonal elements are generally called 'importances' of the corresponding
inverted layer parameters. If the resolution matrix is the identity matrix (1's along the main
diagonal and 0's elsewhere), then each model parameter is uniquely resolved. In practice, we
just desire R to be diagonally dominated.
Joint-inversion
When inverting two (or more) independent EM data sets simultaneously at the same sounding
center to obtain one realistic model, different resolution properties of these data sets are
normally combined. Theoretically, each of them can compensate for the weaknesses of the
other, and therefore the overall benefit would be a reduction of the interpretation ambiguity
[Vozoff and Jupp, 1975; Raiche et al., 1985]. In the 1D layered-earth inversion, the data vector
has the length of both data sets, while the model parameter vector remains of the same length
as for ordinary individual inversion.
With this optimization algorithm the least-squares solution of the EM inverse problem is
regularized to minimize the change of model update or roughness (converse of smoothness)
[Constable et al., 1987]. The philosophy of the Occam's inversion scheme is to find the
smoothest resistivity model which matches the measured data to some required level (i.e. data
misfit tolerance). The interpreter determines how well to fit the data, and the inversion
determines how much structure the model requires. The scheme is a standard implementation
of the 1D smoothed-earth inversion. It tries to solve an under-determined EM inverse problem
(i.e. N<M), and hence requires a starting model which has much more number of layers.
Indeed, the computational expense depends linearly on the number of layers. The layer
thicknesses are fixed, while only resistivities are free parameters in the inversion. Layer
resistivities are then adjusted iteratively until the model response is as close as possible to the
measured data, given constraints which keep the resistivity model smooth. A smoothed-earth
19
Chapter 2 INVERSION OF ELECTROMAGNETIC DATA
Dm 2= (m /z)2 z.
(2.19)
In discrete representation the roughness can be given in terms of a simple matrix operations,
where its penalty matrix is
0. . . . . . . . . . . . .0
1 1 0. . . . . . .0
D= 0 1 1 0. . . . . 0 .
. . . . . . . . . . . . . . .
0. . . . . . . .0 1 1
M M
Here the objective function can be written as
(m)= Dm 2+ 1((m)2*(m)2)
= Dm 2+ 1( WdWf(m) 2*(m)2),
(2.20)
where 1 is the 'Lagrange multiplier' which controls both the model roughness and data
misfit. The 1 is usually chosen so that (m)=*(m), where *(m) is the desired data misfit
(or tolerance).
Following Constable et al. [1987], the objective function is minimized by differentiation with
respect to m and rearranged to yield an expression for the model
m=[DTD+(WJ)TWJ]1(WJ)TWd. (2.21)
20