Professional Documents
Culture Documents
Characterization
An Earth Modelling Perspective
P.M. Doyen
Table
of contents
Table of
contents
5
7
21
43
79
2.1 Overview
2.2 Simple Kriging
2.3 Ordinary Kriging
2.4 Kriging with Locally Variable Mean
2.5 External Drift Kriging
2.6 Cokriging
2.7 Factorial Kriging
2.8 Factorial Cokriging
3.1 Overview
3.2 Sequential Gaussian Simulation
3.3 Sequential Simulation with LVM or Collocated Cokriging
3.4 SGS with Non-linear Relationships
3.5 Simulation with Downscaling
3.6 FFT-Moving Average Simulation
3.7 Gradual Deformation
43
43
51
52
55
58
67
74
79
79
83
87
89
101
103
105
139
4.1
4.2
4.3
4.4
4.5
4.6
5.1
5.2
5.3
5.4
5.5
Overview
Bayesian Classification
Sequential Indicator Simulation with Seismic Constraints
Truncated Gaussian Simulation
Boolean Simulation Methods
Multi-Point Statistics Simulation
Overview
Geostatistical Inversion using SGS
Bayesian Stochastic Inversion
Exploiting Geostatistical Inversion Results
Joint Stochastic Inversion of Elastic and Petrophysical Properties
105
105
120
128
133
133
139
139
142
156
163
Table of contents
165
211
Future Directions
229
References
233
Acknowlegements
247
249
251
Index
253
6.1 Introduction
6.2 Monte Carlo Simulation of Rock Physics Templates
6.3 Combining Rock Physics and Spatial Uncertainty
6.4 Uncertainty Propagation using Linearized Analysis
6.5 Rock Physics Inversion using MCS and Bayesian Methods
6.6 Pseudo-Well Generation
6.7 Direct Petrophysical Inversion
7.1 Overview
7.2 Initial Geomodel Construction
7.3 Downscaling of Dynamic Properties
7.4 PEM-Based Calculation of Elastic Properties
7.5 Computation of 4-D Synthetics
7.6 4-D Interpretation
7.7 4-D Seismic History Matching
165
165
170
179
185
190
197
211
213
213
214
220
222
225
Chapter
1 Introduction
to Geostatistics and Earth Modelling
Chapter 1
Introduction to Geostatistics
from
Seismic Data
and Earth Modelling from Seismic
Data
Chapter 1
used to guide the areal interpolation of zone average well data. Today, 2-D mapping has been largely
replaced by 3-D earth modelling and the focus is on integrating seismic-derived information to constrain fine-scale 3-D property models. The limited vertical resolution of inverted seismic data has been
a major obstacle to the widespread use of seismic information in 3-D property modelling. A number
of promising seismic downscaling techniques have been developed recently and will be reviewed in
Chapter 3.
Since its introduction by Haas and Dubrule in 1994, stochastic inversion remains to this day a very
active area of research. This topic is closely linked to geomodelling because stochastic inversion can
be constrained directly with fine-scale log data and delivers high frequency impedance volumes that
can be integrated in 3-D earth models without downscaling. While deterministic inversion remains
the preferred choice in many companies, the application of stochastic inversion has been growing in
recent years. In parallel with deterministic inversion, stochastic inversion has evolved from post-stack
acoustic impedance inversion to pre-stack implementations in which multiple partial angle stacks are
inverted simultaneously to estimate several elastic parameters. The link between deterministic and
stochastic inversion is also better understood, following in particular the work of Buland et al. (2003a)
in a Bayesian setting. In Chapter 5 we discuss recent developments in stochastic inversion and downstream workflows that exploit multiple realizations of inverted elastic properties for lithology prediction and uncertainty quantification.
Earth models are becoming increasingly complex. They are used to represent properties from
several domains: reservoir properties (e.g., facies, porosity and permeability), seismic properties
(e.g., acoustic impedance), mechanical properties (e.g., effective elastic moduli) and dynamic properties (e.g., saturation, and pore pressure). Data integration between the different domains poses
several challenges, including how to establish relationships between data from different domains
and how to merge the data in a common model framework. Geomechanical and 4-D applications
also require a full earth model covering not just the reservoir interval but also the overburden.
Rock physics provides the necessary link for relating data from different domains (for example a
petro-elastic model is used to link seismic velocities to lithology, porosity, saturation and pressure)
while geostatistics provides the tools to propagate uncertainties in the rock physics transforms. In
Chapter 6, we discuss statistical rock physics, which combines together rock physics and geostatistics. We give a number of examples of rock physics-based inversion and stochastic simulation
workflows for seismic pore pressure prediction in a 3-D geomechanical earth model and for seismic
porosity and lithology estimation in 3-D reservoir models. This chapter also compares sequential
inversion workflows, where elastic inversion is followed by a rock physics inversion, with direct
petrophysical inversion, where seismic data are inverted directly for rock properties such as porosity and lithology. The direct petrophysical inversion workflows are becoming more popular because
they guarantee coherency between the earth model and seismic amplitude data, which is not the
case for traditional cascaded workflows. In Chapter 7, we concentrate on 4-D earth modelling and
the integration of dynamic flow simulation results with time-lapse seismic data in the earth model
framework.
Earth modelling with seismic data is a vast subject. A choice had to be made of the topics covered in
a 1-day course. First, the course concentrates on 3-D property modelling from seismic data. Depth conversion and structural uncertainty modelling will not be addressed. Dubrule (2003) has given a very
good introduction to this important topic. Although Chapter 6 introduces some of the key challenges
in 4-D earth modelling, our coverage is limited to simulator-to-seismic workflows, where 4-D synthetics are created from the flow simulator output and compared to real 4-D data. We do not address the
issues of saturation and pressure prediction from time-lapse seismic data, joint inversion of seismic
and production data and 4-D geomechanics because these topics have been covered in other recent
courses (Calvert, 2005; MacBeth, 2007). Furthermore coverage of these extra topics would require an
extended course period.
In the course, we use the terms earth model, reservoir model and geomodel interchangeably to
mean a 3-D numerical representation of subsurface properties that is obtained by integrating multiple
types of data and can be used to perform physical calculations such a flow simulation and that serves
as a base for making reservoir management decisions. A reservoir model may be seen as a part of an
earth model in that an earth model may represent not just the reservoir but also formations above and
below the producing interval.
Fig. 1.2.1
Chapter 1
Fig. 1.2.2
Fig. 1.2.3
Input data
The main input data for the construction of a seismic-constrained earth model are typically the following:
1. Horizon data and fault picks. A number of horizon surfaces, typically specified as regular 2-D
grids are used to define the top and base of the different reservoir intervals. Horizons are usually
obtained from 3-D seismic interpretation and are depth-converted before integration with the geomodel. Extra stratigraphic surfaces may also be defined between key markers identified from the
seismic data. Fault surfaces or fault polygons or poly-lines picked from seismic data are also a key
input for the geomodel construction.
2. Well log data and well tops. Typically, one or multiple log data tables will be stored for each
well containing measured depth, true vertical depth, porosity, lithology, saturation and other
properties, for each sample point along the well trajectories. When available core data may also
be incorporated to calibrate log-derived porosity estimates or calculate a porosity-permeability
transform. Data tables, containing well picks for the top and base of the different zones of interest, are also required.
3. Seismic attribute volumes. Cubes of depth-converted seismic attributes are obviously a key input.
Depth conversion of the seismic volumes may be performed by resampling the data between key
horizons defined in both time and depth domains. A seismic velocity cube may also be used to
constrain the depth conversion process.
Figure 1.2.1 depicts the main inputs in a structurally simple carbonate reservoir from the Middle
East. In this case, log data at several horizontal wells and a cube of inverted acoustic impedance are
two key inputs for the modelling of the reservoir properties.
Seismic Inversion
Which seismic attributes should be used as input for reservoir modelling? Experience shows that best
results are often obtained when using cubes of inverted attributes such as acoustic impedance. There
are two main benefits of using inverted attributes compared to using amplitudes, AVO attributes or
other attributes resulting from simple transforms of amplitude data. First, inversion partially removes
wavelet effects, increases the resolution and provides impedance values that can be directly correlated with the absolute values of the rock properties stored in our 3-D models. This is not the case of
amplitude or AVO attributes which are related to relative changes in reservoir properties across elastic
interfaces. Second, inverted impedances can be directly calibrated at well locations with log-derived
impedance values. This calibration step, performed in the framework and at the scale of the reservoir
model, is often a crucial step in the construction of a seismic-constrained rock property model.
Recent advances in pre-stack inversion make the use of inverted attributes even more attractive.
In the past, use of post-stack inversion results to constrain earth models was limited by the fact that
only acoustic impedance was inverted from seismic data. It was therefore difficult to separate the
effects of porosity, lithology and fluid saturation from a single inverted parameter. Today, full elastic
inversion is routinely performed by simultaneously inverting multiple partial angle stacks. Figure 1.2.2
shows a typical model-based, simultaneous elastic inversion workflow. An initial model is perturbed
iteratively until a good fit is obtained between real seismic amplitudes and synthetic angle stacks calculated by forward modelling with the Zoeppritz equation. In favourable cases, simultaneous inversion
delivers independent estimates of three elastic parameters, such as P- and S-wave velocities, Vp and Vs
(or corresponding impedances Ip and Is) and density, . Other elastic parameters such as Poisson ratio
can also be derived from inversion results. The availability of 2 or 3 inverted attribute cubes is particularly useful to better constrain earth models and reduce ambiguities when predicting rock properties.
Chapter 1
Fig. 1.2.4
10
Fig. 1.2.5
Furthermore, some inversion tools deliver the inverted elastic properties in stratigraphic grids that can
be used directly for reservoir modelling.
In the illustrative workflow described below and in subsequent chapters, we will focus mainly on the
use of inverted attributes to constrain reservoir models. However, in Chapter 6, we will come back to
the issue of constraining earth models directly from seismic amplitude data.
Depth Conversion QC
Before embarking on the construction of the earth model, basic data QC steps are required. For example, we need to check that depth-converted horizons match the tops defined at the wells. This will be
the case if tops have been used to constrain the depth conversion process. If this is not the case, the
horizons need to be flexed locally near the well locations to ensure that input horizons are consistent
with the well tops. Another important data QC is to check the quality of the inversion at the well locations after time-depth conversion, as shown in Figure 1.2.3. This is done by sampling the depth-converted inverted cubes along the well trajectories and displaying inverted impedance values side-by-side
with log impedance. As shown in the example, inverted results (dotted curve) may not align properly
with the log values due to residual errors in the depth conversion. This problem must be rectified, as it
will adversely affect the petro-seismic calibration at later stages of the workflow. Different strategies can
be applied to correct the residual depthing errors, such as flexing horizons near the wells or applying
residual velocity corrections. This process is often iterative and can be QC by displaying cross plots of
log versus inverted impedance before and after depthing corrections (Figure 1.2.3).
At this stage of the workflow, a decision is made as to whether or not the seismic data can be used
to constrain the reservoir model, based on the quality of the seismic-well tie. Depending on the vertical
resolution of the seismic data compared to the thickness of the reservoir zones, a preliminary decision
may also be made about the type of property modelling technique that will applied: for example, with
high resolution inversion results, we may use a 3-D cokriging procedure while in a thin reservoir zone
we may decide to constrain the model using an areal trend map derived from seismic impedance.
Structural Modelling and 3-D Gridding
Having assembled and QC the data, the first step is to build the structural and stratigraphic framework for the earth model. A 3-D stratigraphic grid is a structured grid composed of hexahedral
cells aligned along layers in order to conform to reservoir stratigraphy. When no faults are present,
stratigraphic grids can be readily constructed in each interval by vertical interpolation between top
and base horizons, as shown in Figure 1.2.4. Different layering styles can be used to conform to the
expected reservoir stratigraphy, as depicted in Figure 1.2.5. For example, in the proportional style,
the cell thickness varies proportionally to the thickness of the gridded interval while in the eroded
style, the stratigraphic grid lines are parallel to the bottom horizon and the cells truncate on the top
horizon, which is interpreted as an erosion surface. Careful selection of the stratigraphic grid style is
important, as it controls the way in which the rock properties are interpolated between well locations.
In reservoir modelling, stratigraphic grids are preferred to regular Cartesian grids because they provide a more natural framework for geostatistical interpolation, as depicted in Figure 1.2.6. Instead of
using Euclidian distance in a regular Cartesian grid (top left), interpolation of reservoir properties is
performed using approximate stratigraphic distances defined along the layers of the stratigraphic grid
(top right). In a stratigraphic grid, each cell has two sets of coordinates: Cartesian coordinates (u,v,w)
defined in the original space and pseudo stratigraphic coordinates (i,j,k) corresponding to the grid
cell indices. Figure 1.2.6 depicts the simple mapping between the two coordinate systems. In order to
ensure proper stratigraphic correlation, geostatistical calculations such as variogram analysis and krig-
11
Chapter 1
Fig. 1.2.6
12
Fig. 1.2.7
Fig. 1.2.8
ing are performed using stratigraphic distances defined in the i,j,k space. In practice, an approximate
stratigraphic distance between two points is computed from the number of cells separating the points
in the i, j and k directions and from the average cell size along each direction.
The dimensions of a stratigraphic grid are typically determined by specifying the size of individual
grid blocks. Typically, grid cells may be 100m x 100m laterally and 1 or 2 metres thick. However grid
block size may vary considerably from field to field depending on reservoir heterogeneity and model
size limitations. In practice, reservoir models will typically contain several millions of grid cells. It is
possible to vary the grid block size in each reservoir interval as shown in Figure 1.2.4. This may be
desirable if we have multiple stacked reservoirs with different levels of spatial variability.
In the presence of faults, stratigraphic gridding is far more complex. Nevertheless, recent
advances in gridding technology and greater commercial availability of sophisticated 3-D geomodelling software have made feasible the construction of faulted stratigraphic grids, such as the one
shown in Figure 1.2.7. Faulted stratigraphic grids have some limitations. For example as shown in
Figure 1.2.7, cells and grid lines are often distorted in the vicinity of faults where columns of cells are
made to conform to the fault geometry. Those distortions may result in biased geostatistical calculations. In Figure 1.2.8, the real stratigraphic distances h1 and h2 are different but they correspond to
the same number of cells. The distances will therefore be wrongly assumed to be identical in variogram calculations made from (i,j,k) coordinate differences. Mallet et al. (2004) have proposed a
new approach where a stratigraphic coordinate system is established from the structural framework
without building any stratigraphic grid. This parametric coordinate system can be used for example
to populate unstructured grids. Nevertheless, experience with geostatistical property modelling on
13
Chapter 1
Fig. 1.2.9
unstructured grids remains limited (e.g., Deutsch et al., 2002a). We will therefore focus exclusively
on reservoir modelling using stratigraphic grids, which is the most widely used approach in commercial applications.
Log Averaging
Log data need to be resampled into the stratigraphic grids in preparation for geostatistical interpolation. In practice, this step is performed by identifying the grid cells that intersect the well trajectories
and averaging the log samples falling within each cell, as shown in the top picture in Figure 1.2.4. For
example, if the log sampling interval is 0.5 ft and the vertical thickness of a grid block is 3 ft, 6 samples
will typically be averaged in each cell along a vertical well. For horizontal wells (Figure 1.2.9), the lateral averaging may include a large number of log samples. The type of averaging operation depends
on the property that is averaged in the grid blocks. Simple arithmetic averaging is used for properties
such as porosity and volume of clay while Backus average may be used to upscale the sonic velocity
logs. Minimum and maximum cut-off values may be introduced for each log to exclude outliers from
the computed averages. Discrete variables such as facies are usually averaged using a majority voting
calculation, which returns as average the most popular facies, based on the relative count of samples
falling in each cell.
Quality control of the log averaging step is performed by comparing averaged or blocked logs
and original logs as shown on the left in Figure 1.2.10. In practice, we examine each well individually to verify that the 3-D model is fine enough vertically to preserve the observed geological layering
and capture the main vertical heterogeneities. If the thickness of the grid blocks in one interval is too
coarse to preserve the required vertical details, the grid is rebuilt with finer cells.
14
Fig. 1.2.10
An important consideration before averaging the logs is to confirm that the 3-D grid honours
exactly the well tops. This is also illustrated in Figure 1.2.10. If this is not the case, data leakage
may occur at the boundary between two geological formations and result in erroneous cell averages.
In practice, this problem is removed by flexing the 3-D stratigraphic grids at well locations before
averaging to ensure that well tops fall exactly on grid-cell facets. This step is particularly important if
an abrupt discontinuity in the logs occurs at the top of the gridded interval.
At this stage of the workflow, another important check is to compare basic statistics such as mean,
variance and coefficient of correlation for different rock properties before and after averaging the
logs in the cells of the stratigraphic grid. Figure 1.2.11 illustrates this analysis using an example
from a carbonate reservoir. Log data from a number of vertical and deviated wells (a) have been
averaged into a stratigraphic grid covering the reservoir interval (b). The cross plot in (c) shows the
impact of the averaging process on the relationship between porosity and acoustic impedance. We
observe that the variance of the grid block data is smaller than that of the log data. This well-known
support effect results from the spatial averaging of log data points in each grid cell. This reduction of
variance leads to an increase in the coefficient of correlation between porosity and impedance. The
level of variance reduction depends on the size of the averaging volume and on the spatial correlation between data values, as explained for example in Journel and Huijbregts (1978). The scaledependence of the variance and of the variogram has an important practical implication: we must
construct histograms and variograms at the correct scale corresponding to the size of grid cells in
our reservoir model. For example, it would not make sense to use the variance of porosity inferred at
log scale as the expected variance level in stochastic simulation done on a coarse 3-D grid. Instead,
we need to use the smaller variance measured after averaging the logs in the stratigraphic grid.
15
Chapter 1
Fig. 1.2.11
16
Fig. 1.2.12
Fig. 1.2.13
A difficult issue is the representativity of the cell average data. We assume that the average of the log
samples is representative of the average property over the volume of the cell. In the example shown
in Figure 1.2.11, we are mixing cell averages computed from vertical and horizontal wells, which may
both provide biased estimates of the true volumetric averages and are computed from widely different
number of samples. Frykman and Deutsch (2002) provide an excellent introductory discussion of the
application of geostatistical scaling laws in reservoir modelling and give a useful list of technical references for this often overlooked topic. Some of the potential pitfalls associated with log averaging are
also discussed in Ringrose (2007).
Seismic Resampling
Depth-converted cubes of inverted seismic attributes are also resampled into the stratigraphic reservoir grid, as illustrated in Figure 1.2.12. The lateral dimensions of individual grid cells (e.g., 50m or
100m) in a stratigraphic grid are typically larger than the seismic CDP spacing (e.g., 12.5m or 25m).
This means that several seismic samples from the original, regularly sampled seismic cube will be averaged laterally into one cell of the stratigraphic grid. Vertically, the situation is just the opposite: the
sampling interval of seismic attribute data in depth will typically be coarser (e.g., 3m to 12 m) than the
thickness of the reservoir grid blocks (e.g., 1 to 2 m). Vertically, the resampling of the seismic cube is
done by interpolating the seismic samples at each point corresponding to a cell centre. Figure 1.2.12
illustrates the resampling of an inverted acoustic impedance cube in the same stratigraphic grid as the
one used in Figure 1.2.9 to average well log data.
The fact that the seismic grid is laterally denser than the stratigraphic grid can give rise to
problems when transferring seismic attributes in the geomodel, as illustrated in Figure 1.2.13 from
17
Chapter 1
Fig. 1.2.14
18
Fig. 1.2.15
Fig. 1.2.16
Hadj-Kacem and Pivot (2005). First, the stratigraphic grid may be mis-positioned with respect to
interpreted seismic horizons and faults (top of figure). This can create artefacts in the seismic property transfer. Second, seismic averaging in the laterally coarser reservoir grid may over-smooth the
heterogeneities captured by the seismic attributes and modify seismic connectivity patterns (bottom
of figure). Hadj-Kacem and Pivot (see also Duplantier et al., 2006 and Hadj-Kacem et al., 2006)
have proposed a workflow combining grid refinement, grid deformation and optimized upscaling
to alleviate some of these problems (Figure 1.2.14). Starting from an original stratigraphic grid (a),
they create a mother grid (b) by lateral refinement of the original reservoir grid. The mother grid
is flexed using a warping technique to better match interpreted horizons (c). Seismic attributes are
then resampled in the mother grid (d) before final upscaling in the coarser reservoir grid (e). They
have also proposed a refined upscaling technique that better preserves seismic facies connectivity
after lateral averaging in the reservoir grid.
Petro-Seismic Calibration
Once both seismic and log data have been transferred into the stratigraphic grid framework, the
petro-seismic calibration can be performed at the scale of the grid cells. If the number of well data
is sufficient, this analysis is conducted separately for each reservoir interval. Figure 1.2.15 illustrates
the petro-seismic calibration for the example introduced in Figure 1.2.11. A cube of inverted acoustic impedance (a) was resampled into the reservoir stratigraphic grid (b). Average porosity can then
be cross plotted against average seismic impedance (d) for all grid cells intersected by wells (c). The
porosity correlation with seismic impedance (orange points) is weaker than its correlation with averaged log impedance (blue points), due to the imperfect estimation of impedance by seismic inversion. The cross plot analysis is an important step, as it determines the strength of the relationship
19
Chapter 1
Fig. 1.2.17
Fig. 1.2.18
20
between seismic impedance and porosity (or other reservoir properties). This analysis is used to define
the statistical correlation parameters for geostatistical interpolation techniques such as cokriging (see
Chapter 2).
Having the seismic data available in a stratigraphic grid is also useful for fast exploration data
analysis. Figure 1.2.16 shows an example of dynamic graphics analysis. A seismic impedance histogram is constructed after transferring the seismic attribute in a stratigraphic grid. The data range (low
impedances) highlighted in the histogram is cross-referenced to the 3-D display where only grid cells
in that range are visible. As expected, low impedance anomalies tend to coincide with the position of
high porosity horizontal wells.
Spatial Continuity Analysis
The next step in the reservoir modelling workflow is to conduct a variogram analysis to characterize
the spatial continuity in the variation of reservoir properties. Variogram calculations are performed
from the well cell data as a function of stratigraphic distances defined directly from the (i,j,k) indices
of the 3-D grids. Distances measured in terms of cell indices are transformed to approximate physical
distances by multiplication with distance increments corresponding to average cell dimensions along
each axis. Figure 1.2.17 depicts an omni-directional variogram calculated as a function of horizontal
(along strata) distance from a number of horizontal wells. Determination of horizontal variograms
is often much harder if the study area contains only widely spaced vertical wells. The concept of
variograms is discussed in the next section.
3-D Seismic-Guided Interpolation
The last step in the construction of a seismic-constrained geomodel involves the interpolation
of reservoir properties between the wells: the geo-cellular framework is in-filled with estimated
or stochastically simulated rock properties. The seismic attribute data are used to guide the
interpolation process away from the wells, as illustrated schematically in Figure 1.2.18. Different
geostatistical techniques exist for this purpose. The most widely used estimation and conditional
simulation techniques will be presented in Chapters 2-3 and 4. Perhaps the most popular method
is cokriging, where seismic and well data are combined in a generalized regression equation.
As explained earlier, the seismic-guided interpolation process is performed in stratigraphic space
after transferring well and seismic data to the reservoir grid.
21
Chapter 1
Fig. 1.3.1
Random Variable and Random Field
A Random Variable (RV) is a variable taking a range of possible values according to a Probability
Distribution Function or PDF. We can distinguish two types of RVs: continuous variables, which take a
continuous range of possible values and discrete or categorical variables, which take only a finite set of
values. The concept of the random variable plays a central role in geostatistical reservoir modelling.
For example, we interpret the value of porosity at one point in the reservoir as a realization, x of a
random variable X. It is common to use upper case letters to represent RVs and lower case letters for
individual realizations. In the course, we will use lower case notation for both, as the context should
make it clear whether we are speaking of the RV or of a realization from it. In geostatistics, we define
RVs at each point in space or in each cell of our reservoir models. We therefore write x(u) or z(u) to
denote RVs representing for example porosity and acoustic impedance at location u. A random field
is simply a collection of RVs indexed by their spatial position.
Why do we use random field theory to describe the reservoir? Spatial variations in reservoir properties and inter-dependence between different properties are complex and cannot usually be described
using simple deterministic functions. Geostatistics provides a framework to represent complex reservoir heterogeneities and model complicated relationships between reservoir attributes. Another reason why we use a statistical framework is that we have incomplete information about the reservoir. If
reservoir properties were measured at every point, we could in principle describe it in a purely deterministic manner. In practice, boreholes sample the reservoir at a very small number of locations. The
task of predicting the properties of a spatially heterogeneous reservoir is therefore highly uncertain.
The probabilistic framework is ideally suited to model this uncertainty.
22
Fig. 1.3.2
Mean, Variance and Gaussian Distribution
Gaussian RVs play an important role in geostatistics. The Gaussian PDF is represented by a symmetric, bell-shaped curve as shown in Figure 1.3.1. It is fully characterized by two parameters: the mean
and the variance. The mean measures the location of the peak of the distribution and the variance
measures the spread of values around the mean. For simplicity, we often write f(x)~N(mx, 2x) to indicate
that x is a Gaussian RV with mean mx and variance 2x. The standard normal distribution is denoted by
N(0,1) to indicate that it has zero mean and unit variance. It can be obtained from another Gaussian
variable x with distribution N(mx, 2x) by applying the change of variable (xmx)/x. Figure 1.3.1 shows
two Gaussian distributions with different means and variances. The concepts of mean (or expected
value) and variance apply to any distribution whether it is Gaussian or not. In practice, sample means
and variances are calculated using standard formulae.
Cumulative Distribution Function
Figure 1.3.2 gives the definition of the Cumulative Distribution Function (CDF) and shows the CDF
corresponding to a Gaussian PDF. The CDF for one value x is obtained by integrating the corresponding PDF over all values below x. It is particularly useful to calculate an interval probability or a probability of exceedence. Figure 1.3.3 shows the 95% confidence interval for a Gaussian PDF. It means that
if we draw values at random from the PDF, 95% of the values will fall in the interval [m2,m+2].
Linear Correlation and Bi-variate Gaussian Distribution
In the course, we will often relate two or more RVs at the same location, for example when we want
to predict porosity, x from acoustic impedance z. The relationship between the two variables is usually
23
Chapter 1
Fig. 1.3.3
Fig. 1.3.4
24
Fig. 1.3.5
summarized using their covariance or coefficient of correlation. This concept is explained in Figure
1.3.4. The coefficient of correlation, xz=, is just a dimensionless, normalized covariance obtained
by dividing the covariance by the product of the standard deviations of the two variables. The coefficient of correlation measures the degree of linear dependence between two variables and varies from
-1 (perfect correlation with negative slope) to +1 (perfect correlation with positive slope). A value of
zero means that there is no linear relationship between the two variables. Kalkomey (1997) provides a
useful discussion of the pitfalls involved in linear correlation analysis: when samples used to estimate
are sparse, it is very easy to obtain spurious high correlation values. Non-linear relationships may
also exist between two attributes that cannot be measured by . We will come back to this situation
in Chapter 3, where we discuss strategies for handling non-linear relationships between seismic and
reservoir properties in stochastic simulations.
The concepts of covariance and correlation coefficient are strongly linked with Gaussian variables.
Figure 1.3.5 defines the bi-variate Gaussian distribution, f(z,x)=f(z). We use bold letters to denote vectors. The joint Gaussian PDF is fully specified by the mean vector, which corresponds to the means of
the two RVs and the covariance matrix, which is constructed from the variances of the two variables
and their covariance. The covariance matrix is symmetrical. This means that the off-diagonal elements
are equal and can be expressed in terms of the coefficient of correlation as Cxz=Czx=xz. The same
exponential expression applies for multivariate Gaussian distribution of any dimension. In n dimensions, the covariance matrix has elements Cij=E[(xi mi)(xj mj)], which correspond to the covariances
between variables xi and xj with means mi and mj, for i and j = 1, , n. Figure 1.3.5 shows iso-probability
contours of the bi-variate Gaussian PDF. They conform to ellipses centred at the mean location mz,
which corresponds also to the PDF maximum. Figure 1.3.5 also shows samples drawn at random from
25
Chapter 1
Fig. 1.3.6
Fig. 1.3.7
26
the joint PDF. They define a cloud of points centred at the mean point. The shape of the data cloud
depends on the elongation and orientation of the elliptical iso-probability contours, which depend
themselves on the covariance matrix, as shown in Figure 1.3.4. When RVs x and z are uncorrelated (i.e.,
=0), lines of equal probability become circles. Integrating the joint PDF over all values of z yields the
univariate Normal distribution of x, N(mx, 2x), as shown in the top left corner of Figure 1.3.5.
Gaussian Conditional Distribution and Linear Prediction
We introduce the concept of Linear Mean Square (LMS) prediction in a simple Bayesian context.
Suppose that we have two correlated Gaussian RVs z and x with correlation coefficient . The variables
represent for example acoustic impedance and porosity at the same location. Suppose also that we
have measured the value of the acoustic impedance z. We would like to make a prediction of x based
on the measured value of z using the fact that the two variables are linearly correlated. We will denote
our prediction of x by x.
First, we ignore the z measurement. In that case, we predict x by the mean mx, i.e., x = mx. It is easy to
show that this value would be the best prediction in that it minimizes the Mean Square Error (MSE),
E(xx)2. Furthermore, the MSE for this best prediction is equal to the variance of the PDF of x, 2x. In
the absence of z data, we may see the PDF f(x) as a prior distribution whose mean is our best estimate
of the unknown value of x and whose variance measures the prediction uncertainty.
Next, we would like to update our best estimate of x using the z observation. If the two RVs are correlated, we expect that the range of possible values of x corresponding to the measured z will be narrower
than the full range of values associated with the prior PDF f(x). This restricted range of x values corresponds to the spread of the bi-variate distribution along a vertical line drawn in the cross-plot of Figure
1.3.6. Statistically, we represent this updated range of values by the conditional distribution f(x|z). This
distribution is also called the posterior distribution in that it represents the range of possible outcomes for
x updated from the knowledge of z. If the bi-variate distribution is Gaussian, the conditional distribution
is also Gaussian with conditional mean and variance as given in Figure 1.3.6. Measurement of z leads to
a reduction of the variance that is directly related to the magnitude of the coefficient of correlation. If
=1, the posterior has zero variance; if =0, the two variables are uncorrelated and the posterior variance is equal to the prior variance 2x. In this Gaussian setting, the posterior variance does not depend
on the value of z but the posterior (or conditional) mean is a linear function of z. The posterior mean
also corresponds to the Maximum A Posteriori (MAP) value of x, i.e., the value of x corresponding to the
maximum of the Gaussian PDF f(x|z). In a Bayesian setting, the posterior mean would also be used as
our best estimate of x given the z data point.
The same results can be obtained using standard LMS prediction theory without reference to Gaussian
variables: we construct a linear estimate of x as x=a+bz and we obtain the values of a and b by minimizing the MS prediction error E(xx
)2 with the extra condition that the estimate be unbiased, i.e., E(x
)=mx.
The resulting linear regression is identical to the conditional mean formula and the predicted MSE is
the same as the conditional variance. Nevertheless, the Bayesian framework is more general than least
squares. For example, we can still use the concept of posterior distribution when the RVs z and x are nonlinearly correlated. This will be discussed in the context of stochastic simulation in Chapter 3.
Spatial Correlation and Linear Prediction
Spatial correlation is a key concept in geostatistics. It allows us to make prediction at one point in the
reservoir from observations made at different locations. Figure 1.3.7 shows three synthetic porosity profiles, representing measurements along three hypothetical horizontal wells. The experimental mean and
27
Chapter 1
Spatial Correlation
Fig. 1.3.8
Fig. 1.3.9
28
Fig. 1.3.10
variance of porosity are exactly the same in the three wells. In the absence of any extra information, we
would wrongly conclude that the uncertainty in predicting porosity would be the same in all three cases
because the spread of values is the same. Of course, we know intuitively that it will be much easier to
predict porosity away from the well in case A than in case C because the lateral variations are much less
erratic. It is clear that univariate statistics do not allow us to differentiate the three different scenarios of
reservoir heterogeneity. Instead, we need to introduce the concept of spatial correlation in order to quantify the various degrees of lateral variability. For this purpose, we first introduce the concept of h-scatterplot (e.g., Isaaks and Srivastava, 1989) in Figure 1.3.8. We use x(u) to denote the value of porosity at one
point of coordinate u along the horizontal well and x(u+h) to represent the porosity value at distance h
away from the first point. A h-scatterplot is constructed by cross-plotting x(u+h) versus x(u) for all pairs of
points separated by distance h. Figure 1.3.8 shows a series of such h-scatterplots for increasing distance h.
We can compute the coefficient of correlation between cross-plotted variables for each lag distance h. The
cross-plots become more diffuse as h increases because the porosity values become less and less similar.
An experimental spatial correlation function is obtained by plotting the coefficient of correlation as a
function of h, as shown in Figure1.3.9. (h) measures the degree of linear correlation between porosity values as a function of interdistance h. The plot of (h) shows the expected decrease in the level of
correlation with increasing h. The distance at which the correlation becomes zero is called the range of
correlation or the correlation length. In this case, the lateral correlation range is about 50 m. Beyond this
distance, porosity values are approximately uncorrelated. Figure 1.3.9 also gives the formulae defining
the theoretical spatial correlation and covariance functions and the corresponding expressions for calculating them experimentally. The theoretical formulae for C(h) and (h) are identical to the formulae
shown in Figure 1.3.4, except that z and x have been replaced by x(u) and x(u+h), respectively. We have
used the superscript * to differentiate the experimental formulae from the theoretical ones.
29
Chapter 1
Fig. 1.3.11
Figure 1.3.10, we plot the spatial correlation functions corresponding to the three horizontal wells.
The lateral continuity in well A is much greater than in well C. This translates into a much larger
correlation length, as expected. The correlation range controls the degree of predictability of porosity away from the well. Suppose, we wish to predict porosity x(u+h) at a distance h from one porosity
measurement x(u). This problem is conceptually similar to the problem of predicting porosity from
acoustic impedance at the same location illustrated earlier in Figure 1.3.6. The only difference is that
instead of using the local correlation between z and x, we use the spatial correlation between x(u) and
x(u+h) to build a best linear estimate of porosity, as shown in Figure 1.3.11. The expressions for the
conditional expectation and conditional variance are the same as before but we have replaced the local
coefficient of correlation zx by the spatial correlation (h). The formula for the conditional expectation makes it clear that (h) measures the degree of linear dependence between porosity values as a
function of distance. If we predict porosity at a distance h greater than the correlation range, the best
linear porosity estimate becomes equal to the porosity mean mx and the predicted mean square error
(or the estimation variance) becomes equal to the a priori variance. In Chapter 2, we will come back
to this example and show that it is a special case of linear estimation with Simple Kriging.
Stationarity
When constructing the h-scatterplots and the experimental spatial correlation function displayed in
Figures 1.3.8 and 1.3.9, we have made implicitly the assumption that we could pool all the well data
together and calculate meaningful summary statistics. In fact, we have made the two assumptions of
stationarity listed in Figure 1.3.12. First, we assume that the means of the RVs x(u) and x(u+h) do not
vary along the section of the well trajectory under investigation. This allows us to estimate mx as the
sample arithmetic mean of all the data values. Second, we assume that the covariance between RVs x(u)
30
Stationarity
Fig. 1.3.12
and x(u+h) depends only on the lag distance h. This allows us to compute the experimental correlation, *(h) at distance h by combining all pairs of points separated by that distance. Stationarity of the
covariance implies that the variance 2x = C(0) is also stationary and can be estimated as the sample
variance of all the data. Figure 1.3.12 shows two examples where these assumptions are violated. In
example (a), the porosity increases systematically from left to right, implying that the mean is spatially
variable, i.e., mx = mx(u). In example (b), the level of local variability increases from left to right with
increasing porosity, implying that both the mean and variance depend on the position u. The first
non-stationary situation is common in practice and corresponds for example to the existence of a
geological trend. To tackle this situation, we will often decompose the spatial variations of x into two
components: a smoothly varying deterministic function corresponding to the trend and a more rapidly
varying random component called the residual. This will be discussed further in Chapter 2 (see Figure
2.4.1), where we introduce the concept of kriging with a locally variable mean.
Variogram
Geophysicists are more familiar with correlation or covariance functions but geostatisticians tend to
prefer variograms (Figure 1.3.13) because their computation does not require estimation of the mean.
Furthermore, the variogram is somewhat more general as its definition requires the weaker stationarity
assumption that the increments x(u+h)-x(u) be stationary. This means that variogram analysis can be
applied to random phenomena which have unbounded variance. However in most practical situations,
the two tools are equivalent. Figure 1.3.13 gives the formula defining the variogram as the variance of
the increments. In practice, the experimental variogram is constructed by plotting the average square
difference between data values as a function of distance h. For each lag distance h, the average is taken
over all pairs of points separated by that distance. The variogram is related to the covariance by the
31
Chapter 1
Fig. 1.3.13
Courtesy of T. Colou
32
Fig. 1.3.14
Modelling Anisotropy
Fig. 1.3.15
simple formula shown on the left in Figure 1.3.13. The variogram tends to increase with increasing
distance because it measures the dissimilarity between data values as a function of h. The opposite is
true for the covariance because it measures the level of similarity as a function of h. Figure 1.3.14 summarizes key variogram parameters:
1. Variogram of stationary RVs reach a plateau level at a certain distance h. That distance r corresponds to the correlation range or correlation length. Beyond that distance, data values are uncorrelated.
2. The level of the plateau is called the sill of the variogram. It corresponds to the level of variance of
the data. The higher the sill, the greater the variability.
3. Variograms often exhibit a discontinuity at the origin (h=0). This is called the nugget effect in
geostatistics. It corresponds either to pure random noise or to variability at a scale smaller than the
sampling interval.
Directional Correlogram and 2-D Correlation Map
So far we have computed a spatial correlation or variogram function along one direction, corresponding to the orientation of a horizontal well in our example. In general because of anisotropy,
the degree of lateral continuity in a reservoir will be directionally dependent. We therefore compute
a series of 1-D experimental correlograms or variograms along different directions. Figure 1.3.15
illustrates the concept of directional correlogram using a seismic amplitude map (a) from a North
Sea reservoir. In (b), we show a series of correlograms computed for different azimuths . In polar
coordinates, we can represent a distance vector (hu,hv) by (h, ) where h is the length of the distance
vector aligned along direction . Each correlogram measures the degree of correlation between
33
Chapter 1
Fig. 1.3.16
Fig. 1.3.17
34
seismic amplitude values with increasing distance h measured along a particular direction. The correlation ranges have been interpreted manually along each direction and are displayed as coloured
sticks in (c) radiating from a common origin corresponding to a zero separation distance. Often the
experimental ranges will approximately fall on an ellipse, as shown in (c). This situation is called
geometric anisotropy and is characterized by two principal correlation ranges, corresponding to the
half-length of the major and minor axis of the ellipse. The orientation of the ellipse is measured
by the azimuth of the major axis, which corresponds to the direction of greatest lateral continuity.
When data are regularly spaced in 2-D, we can also compute a correlation map as shown in (e).
Directional correlograms represent radial slices passing through the origin of the map. For example, the correlogram shown in the top right corner in light green colour corresponds to a slice with
azimuth =45o. Figure 1.3.16 shows another example of a 2-D spatial correlation map (b) computed
from the 2-D grid of data displayed in (a). The figure also explains how the experimental covariance and correlogram are computed. For a distance vector h=(hu,hv), the experimental covariance
is calculated by averaging the product of data values over all pairs of grid points separated by h. In
practice, the covariance is computed for distance lags hu and hv corresponding to integer multiples
(k and l) of the grid node spacing . The covariance is a symmetric function, i.e., C(h)=C(-h), as
seen from the map display. When data are irregularly distributed in 2-D, directional correlograms
or variograms can still be computed in a similar manner. The procedure, which involves binning
pairs of data points according to distance and angle classes, will not be described here. Details can
be found for example in Deutsch (2002).
Modelling 2-D Correlograms or Variograms
Except in special circumstances (see for example Section 2.8 on automatic factorial cokriging),
experimental correlation values are not used directly. Instead, parametric models are fitted to the
experimental correlograms or variograms. There are two main reasons why modelling is performed:
W
hen applying kriging or simulation, covariances may have to be evaluated for distances not
available from sample covariances. There is therefore a need to interpolate between experimental
covariance values.
Kriging matrices (see Chapter 2) constructed from covariance values must be positive definite to be
invertible. The parametric models fitted to the experimental covariances ensure that this condition
is respected.
In practice, variogram modelling is performed from a set of valid parametric models such as the
popular isotropic covariance models shown in Figure 1.3.17. We use the notation C(h/a) to make clear
that the correlation range acts as a scale factor on the modulus h of the distance vector h. The Gaussian
and exponential models reach zero asymptotically when h goes to infinity. However, a practical correlation range is defined as the distance when the covariance reaches 5% of the value at the origin. A
factor of 3 has been introduced in the two exponential formulae so that their practical range of correlation is equal to a.
Anisotropic covariance models are constructed from isotropic functions. As discussed earlier, we
consider the situation where plots of directional covariance as a function of h have the same shape but
different ranges of correlations. Furthermore, we assume that the azimuthal dependence of the range
can be described using an ellipse. Figure 1.3.18 (bottom) gives the equation of the ellipse radius, re as a
function of for an ellipse parameterized using the (semi) major and minor axis lengths, a and b, and
the orientation of the major axis. (The azimuth is usually measured clockwise from the N-S (hv) axis,
so in this example is negative.) The anisotropic covariance function is obtained from an isotropic
35
Chapter 1
Fig. 1.3.18
Fig. 1.3.19
36
Fig. 1.3.20
model by scaling the range according to the azimuthal dependence of re. Figure 1.3.19 shows two 2-D
anisotropic exponential covariance models obtained by scaling the isotropic model shown on the left.
Experimental spatial covariances may reveal several scales of variability, each with its own anisotropy
and correlation ranges. The experimental correlograms of the seismic amplitude map in Figure 1.3.15
clearly capture two scales of variability: a small-scale structure corresponding to the rapid decrease of the
correlation level at short distances and a longer-scale structure corresponding to the more gradual decay of
the correlograms at intermediate distances. The two nested structures visible in the correlograms are easily
modelled using a sum of two covariance models: one of them has already been discussed; it corresponds
to a geometrically anisotropic model with correlation ranges and orientation shown in (c). The small-scale
structure has been modelled using an isotropic model with range equal to 400m (d). The combined fitted
model is overlaid on the experimental correlograms in (b). The relative variance of each model component
has been scaled so that the sum of the two components add to one at the origin (h=0).
3-D Variograms and Correlograms
Directional variograms or correlograms can also be calculated in 3-D. In this case x(u) represents for
example the porosity at location u=(u,v,w) and x(u+h) represents porosity at another location separated by the distance vector h=(hu ,hv ,hw). Experimental variograms are computed as a function of distance |h| along different directions specified by two angles as shown in Figure 1.3.20: the azimuth and
dip angles. Interactive software tools are typically used to perform the variogram analysis, as shown
in Figure 1.3.21. A stereo net is used to specify the dips and azimuths for directional variogram computation. Variogram ranges are picked manually along each direction and are displayed as a 3-D stick
diagram. The 2-D ellipse of anisotropy is replaced by a 3-D ellipsoid, fully specified by three angles
37
Chapter 1
Fig. 1.3.21
Fig. 1.3.22
38
and three principal correlation ranges, corresponding to the three axis of the ellipsoid. In layered
reservoirs and when working in stratigraphic coordinates, we expect the minor axes of anisotropy to
be vertical (perpendicular to bedding) and the medium and major axes to be horizontal (or parallel
to the layering). Experimental variograms are therefore computed in various horizontal directions
(dip = 0o) and in the vertical direction (dip = 90o). In the stereo-net displayed in Figure 1.3.21, the vertical direction corresponds to a dot at the centre of the net whereas horizontal directions correspond
to dots located on the perimeter the colored circle.
Figure 1.3.22 from Caers (2005) depicts the 3-D porosity distribution in a set of synthetic reservoir
layers and the corresponding directional variograms. They provide a good illustration of the impact
of the variogram structure on 3-D spatial continuity. In particular:
1. All variograms increase with increasing distance to reach a plateau level corresponding to the sill,
except in (d) where the lateral correlation length is large compared to the size of the 3-D block and
corresponds to a laterally homogeneous reservoir.
2. As expected, the vertical correlation ranges are much smaller than the horizontal ranges, indicating
that there is more continuity parallel to layering than in the perpendicular direction.
3. The slope of the N-S horizontal variogram at the origin is greater in reservoir (b) than in reservoir
(a), meaning that the variogram increases more quickly for small distance h in (b) than in (a). The
behaviour of the variogram near the origin controls the smoothness of the porosity variations: the
gentler variogram slope in (a) gives rise to smoother spatial variations.
4. The N-S variogram in (c) exhibits a strong apparent discontinuity at the origin. This nugget effect
translates into more erratic variations, representing variability at a scale smaller than the size of the
grid blocks.
39
Chapter 1
Fig. 1.3.23
Spatial Cross-Correlation
Fig. 1.3.24
40
Fig. 1.3.25
The main difficulty in variogram analysis is often the lack of data to compute reliable statistics.
The vertical variogram structure is typically well-defined from log data at vertical wells. Except when
horizontal wells are available (Figure 1.2.17), it is often impossible to define the horizontal variogram
structure because the well spacing is about equal or larger than the lateral correlation range. Figure
1.3.23 from Cole et al. (2003) illustrates a strategy that can be employed in such situations. After resampling in the reservoir stratigraphic grid, porosity log data from 27 wells (a) were used to compute a
vertical variogram with range of about 50feet (b). Because of the sparse nature of the well data (average
well spacing is several kilometres), not enough data pairs are available to compute reliable directional
variograms in the horizontal direction (i.e., horizontal in stratigraphic coordinate space). Instead, an
omni-directional horizontal porosity variogram was computed from the well data. This means that the
variogram calculation is averaged over all directions to obtain more data pairs and a less noisy variogram estimate. The omni-directional experimental variogram (c) is still noisy but a spherical model
with range of about 7km provides a reasonable fit. In order to study the lateral anisotropy, directional
correlograms were obtained from a seismic impedance map (d), representing average values over
the reservoir interval. Horizontal directional variograms were computed (e) and fitted using a geometrically anisotropic spherical model with principal correlation ranges and azimuth as shown on
the correlation map (f). Finally, the lateral anisotropy model for acoustic impedance is assumed to be
representative of the anisotropy of the porosity variations. The 3-D ellipse of anisotropy for porosity (g) is therefore constructed by combining the vertical range calculated from the well data and the
horizontal ranges computed from the seismic data. This model is also validated because it is consistent
with the preferential directions of lateral continuity expected in this carbonate depositional setting.
Furthermore, the omni-directional range value of 7km calculated from the well data is in between the
major and minor ranges computed from seismic impedance, as it should be.
41
Chapter 1
Spatial Cross-Correlation
Earlier in this chapter, we introduced the local coefficient of correlation between two attributes such as
acoustic impedance and porosity. This concept can be extended to model the spatial cross-correlation
between two attributes. For simplicity, we restrict our discussion to a 1-D example. Suppose that we
want to relate lateral variations of porosity to changes in seismic amplitude along a horizontal profile,
as shown in Figure 1.3.24. We use x(u) to denote the value of porosity at one point of coordinate u
and z(u+h) to represent the amplitude value at distance h away from the first point. A h-scatterplot is
constructed by cross-plotting z(u+h) versus x(u) for all pairs of points separated by distance h. Figure
1.3.24 shows a series of such h-scatterplots for increasing distance h. From the scatterplots, we can compute the cross-correlation coefficient between x and z for each lag distance h. We can then construct an
experimental cross-correlation function by plotting the value of xz as a function of h. In general, the
cross-correlation will be maximum at the origin and decrease with increasing distance. The correlation length can be defined as the distance at which the values of z and x become uncorrelated. Figure
1.3.25 gives the definition of the spatial cross-covariance and cross-correlation functions. Cross-covariances cannot be modelled independently of the auto-covariances of the two attributes z and x. Instead,
a linear model of coregionalization must be used (e.g., Journel and Huijbregts, 1978). In practice,
modelling of cross-covariances is quite tedious, especially if more than two variables are involved.
As we will see in the next chapter, most people today use collocated cokriging with a Markov-type
assumption. Under this assumption, the cross-covariance function is supposed to be proportional to
the covariance of the primary attribute. The only parameter to calculate is therefore the coefficient of
correlation, xz (0), which measures the strength of the local linear relationship between the primary
and secondary variables.
42
Chapter
2 3-D Geostatistical Interpolation
& Filtering
3-D Geostatistical
Interpolation & Filtering
2.1 Overview
Kriging-based techniques are widely used for interpolating reservoir properties. Kriging comes in
many different flavours such as simple kriging, universal kriging, cokriging and factorial kriging to
name a few. In this chapter we first review the basic concept behind kriging. Next, instead of giving
a complete catalogue of the different kriging variants, we discuss the kriging-based techniques most
suitable for integrating seismic information in 3-D reservoir modelling. We also focus on the application of kriging for filtering seismic images, as this topic has received significant attention in the last
few years. This topic is also of interest because it highlights the importance of properly conditioning
seismic data before using them for quantitative reservoir characterization.
Fig. 2.2.1
43
Chapter 2
Simple Kriging
Fig. 2.2.2
Fig. 2.2.3
44
Fig. 2.2.4
whereas linear prediction filtering theory was first developed in the context of time series analysis, with
regularly sampled data. The basic idea of kriging is explained in Figure 2.2.1 in the context of predicting an unknown rock property, xo, for example porosity, at one location uo in a 3-D grid. In the Simple
Kriging (SK) approach, we obtain an estimate, x0sk, of the unknown rock property as a weighted linear
combination of the well data values, xi. To simplify notation, we use the index i to refer to a particular
location ui, representing the centre of a 3-D grid cell. In the figure, mx represents the expected value
of variable x. In SK, the mean is assumed to be spatially constant and is fixed by the user. In order to
obtain the kriging estimate, we must calculate the weights, wi, of the different data values. We would
like to select these weights so that we minimize some measure of the unknown prediction error x0 x0sk
at location uo. For kriging, we choose to minimize this error in a mean square sense (Figure 2.2.2). This
amounts to minimizing a quadratic error function parameterized in terms of C, the spatial covariance
function of variable x. In practice, we obtain the weights corresponding to the optimal solution by solving
a system of equations called kriging system, which are also called normal equations in time series analysis.
An example of kriging system is shown in Figure 2.2.3 in a situation where we have three data points.
The elements of the symmetric kriging matrix, Cij, represent spatial covariances evaluated for distance
vectors hij between data points i and j. The diagonal elements of the matrix are all equal to the variance
of x ( 2x = 2). The kriging vector on the right-hand side of the matrix equation is obtained from spatial
covariances calculated for distance vectors between the three data locations and the position uo, where
we make the prediction. The kriging matrix captures information about data correlation or redundancy
while the kriging vector provides spatial correlation information between the estimation location and the
data locations. It is interesting to note that the kriging system depends only on the data configuration
relative to the estimation location and not on the data values themselves.
45
Chapter 2
Fig. 2.2.5
How do we solve the kriging system? In practice we invert the kriging matrix and multiply it by the
kriging vector to obtain the solution for the weight vector. To be invertible the matrix must be positive
definite. This is why we cannot just plug in experimental covariance values in the matrix, as the resulting matrix could become ill-conditioned. Instead, as explained in Chapter 1, we fit a parametric model
to the experimental covariances that guarantees the condition of positive definiteness by construction.
An important property of kriging is that it is an exact interpolator; i.e., it reproduces exactly the data
values, as shown in 3-D in Figure 2.2.4.
Search neighbourhood
How do we determine the data points that are included in the kriging system? For large datasets or for
reasons of stationarity, we often use a moving search neighbourhood. Data points are only retained in
the kriging system if they fall within a local search region defined for example by a 2-D ellipse centred
at the estimation location (Figure 2.2.5) or an ellipsoid in 3-D, with major and minor axes aligned
along the principal directions of spatial anisotropy. To minimize directional interpolation artefacts, the
search template is divided into quadrants or octants and data points are selected so that they provide
omni-directional coverage. When data are clustered, adaptive search strategy can be used: the size
of the search region is increased until a sufficient number of points is found. With a moving search
region, the data configuration changes each time we estimate xo at a new grid point uo. The steps of
building the kriging system and inverting it are therefore repeated at each location in order to find the
optimum set of weights. In that sense, kriging is an adaptive technique in that the weights are adapted
to the local data configuration.
46
Fig. 2.2.6
An alternative strategy is called kriging with a unique neighbourhood: instead of using a moving search
region, we use all available data points in the kriging system at each grid cell / node location. This is
handy because we only need to build and invert the kriging matrix once. The kriging vector (right-hand
side of the matrix equation in Figure 2.2.3) still needs to be recalculated at each predicted location but
kriging with unique neighbourhood can provide an important CPU cost saving for large grids, as most
CPU time is spent performing matrix inversion calculations. Another advantage of the unique neighbourhood strategy is that it minimizes the risk of kriging artefacts: when using a moving search region,
spurious discontinuities occur in kriging results when the number of data points in the search region
changes abruptly from one location to the next. Kriging artefacts with a moving search region can be
particularly severe when extrapolating away from a cluster of well data, as shown in Figure 2.2.5.
In practice, application of the unique neighbourhood technique is limited by the number of control
points (a few hundreds) otherwise the matrix to invert becomes too large and ill-conditioned. In addition to kriging matrix size, the applicability of the unique neighbourhood technique is limited by the
underlying assumption of stationarity and validity of the spatial covariance model at large distances.
For example, in 3-D, we often observe a vertical trend in the reservoir interval. Mixing well data points
from different depths in a simple kriging system is therefore not advisable, as it assumes a spatially
constant mean. A better approach, often used in practice, is to apply unique neighbourhood kriging
layer-by-layer in the 3-D grid, with a layer-dependent constant mean calculated by averaging the well
data at each stratigraphic level.
47
Chapter 2
Fig. 2.2.7
Fig. 2.2.8
Fig. 2.2.9
49
Chapter 2
Fig. 2.2.10
Fig. 2.2.11
50
Fig. 2.2.12
Finally, imagine that we have a single well location and want to predict porosity (x) at a distance h away
from the well (Figure 2.2.11). The kriging estimate reduces to a traditional linear least-squares regression
with a correlation coefficient dependent on the inter-distance between the data point and the estimation
location. The same result was obtained in Chapter 1 (Figure 1.3.11) using conventional correlation analysis.
Kriging Variance
As a by-product, kriging provides a measure of the prediction error, which is called the kriging variance
or the predicted mean square error. It corresponds to the value of the quadratic error surface at the minimum point (Figure 2.2.2), which determines the kriging weight values. Under a multi-Gaussian assumption with known covariance model, the kriging variance provides a confidence interval for the estimated
value (Figure 2.2.12). In practice, it may be difficult to verify this Gaussian assumption. Furthermore, the
kriging variance depends on the selected spatial covariance model, which is also uncertain, especially if
well control is sparse. The kriging variance can therefore not be interpreted, as providing absolute error
bounds on the predicted values. Instead, the kriging variance gives a relative measure of the quality of
the interpolation in different regions of the reservoir and reflects the spatial distribution of the data. As
shown in the 3-D example in Figure 2.2.12, the kriging variance is zero at well locations, as kriging is
an exact interpolator. At distance from wells greater than the correlation range, the kriging variance is
equal to the variance of the rock property.
51
Chapter 2
Ordinary Kriging
Fig. 2.3.1
form of kriging is called Ordinary Kriging. In OK, the mean mx is assumed to be a priori unknown.
The OK estimator is made unbiased by requiring that the kriging weights sum to 1. This is done by
introducing an extra constraint in the kriging system. How does OK relate to SK? As shown schematically in Figure 2.3.1, OK amounts to re-evaluating the mean at each estimation location using the data
falling within the moving search neighbourhood. Simple kriging is then applied using the residuals
from the implicitly estimated mean. In theory, OK is more data-adaptive than SK in that it can track
smooth variations in the mean corresponding for example to a geological trend. In practice however,
OK can give rise to artefacts in the form of abrupt discontinuities in the mapping results. This occurs
when the number of data points contained in the moving search region is sparse, leading to abrupt
variations in the mean recalculated at each point. Another potential problem with OK occurs when the
data provide a biased sampling of the reservoir, for example if all wells are located in a high porosity
zone. In that case it is better to impose a guessed mean value with SK or provide an external trend
model than to use OK with an implicitly biased mean.
52
Fig. 2.4.1
Fig. 2.4.2
53
Chapter 2
Fig. 2.4.3
In practice the trend function m(u) can be specified as a polynomial function of u. For example, a
compaction trend can be modelled using a linear or quadratic depth function fitted by least squares
to well log data.
In the context of seismic-guided prediction of porosity, a useful option is first to convert a cube
of seismic impedance (or another seismic attribute) into porosity using standard regression analysis.
The seismic-derived porosity cube is then used as trend in the kriging with LVM scheme. This simple
technique works quite well in practice. The kriging of the residuals guarantees that the resulting model
honours the well data. Away from wells, the final porosity model reduces to the externally specified
trend, i.e., when all data points are located at a distance greater than the correlation range from the
estimation location uo, the kriging residual is zero. Figure 2.4.3 illustrates this trend-kriging approach
with a seismic porosity prediction case study for a carbonate reservoir from Saudi Arabia (Cole et al.,
2003). In this example, a cross-plot of porosity versus Acoustic Impedance (AI) revealed two trends
related to changes in lithology: the curved trend (in red) corresponds to good quality reservoir rocks
with dolomitic facies while the lower linear trend (in blue) is associated with tight limestone and mudstone. At first, the facies-specific porosity-AI relationships were applied separately, after building a 3-D
facies model. This solution led to unrealistically abrupt changes in porosity across facies boundaries.
The solution adopted next was to define a 3-D porosity trend by linear blending of the two relationships with weights equal to the corresponding facies probabilities. The resulting seismic-derived porosity trend model (in the middle at the bottom of the figure) was combined cell-by-cell with the kriged
well residuals to derive the trend-kriged porosity model (bottom right) that ties at the wells and is consistent with the seismic trend. The advantage of kriging with LVM over more sophisticated techniques
such as kriging with an external drift is that the user has complete control over the trend model and
can incorporate different geological constraints, as the example shows.
54
Fig. 2.5.1
55
Chapter 2
Fig. 2.5.2
Fig. 2.5.3
56
Fig. 2.5.4
polation process. In this case (Figure 2.5.1), our trend model is defined not as a polynomial functions
of the geographic coordinates but as function of the travel time defined at each interpolation grid
node. A simple linear relationship is often used to define the trend. The steps involved in EDK prediction of velocity are described in Figures 2.5.2 to 2.5.4. Our goal is to estimate the unknown velocity at
a grid node (red dot in Figure 2.5.2) at which we only know the interval travel time To. EDK predicts
Vo by combining To with velocity and interval travel time data measured at wells located inside a search
region centred at the estimation location. In the example, we have three wells inside the search region,
represented by the orange circles. In a first step, we estimate the coefficients a and b of the velocitytravel time trend by regression, using only the (Vi, Ti) data points located inside the search region. (We
use and b to denote the estimated trend coefficients). This means that the trend coefficients will be
recalculated at each grid node if we use a moving search neighbourhood. In the second step (Figure
2.5.3), we define the residuals at the wells by subtracting the local value of the trend. Next, we krige the
residuals at the estimation location. Finally, we obtain the EDK estimate of the velocity, V0edk by adding
the trend computed locally to the kriging residual. In practice, this multi-step procedure is performed
implicitly at each grid node by solving an EDK system of equations, as shown in Figure 2.5.4. Like SK,
EDK is an exact interpolator. In extrapolation (when Ci0= 0 for all data points), the EDK prediction
reduces to the implicitly estimated velocity trend, m
0 = +bT0. When using a moving search region, the
trend coefficients are re-estimated at each grid node. This allows in principle to track non-stationarity
in the velocity-time relationship. In practice, this data adaptive feature is of limited value in a reservoir
modelling context because the number of wells is usually too small to obtain stable estimates of the
space-variant regression coefficients. Furthermore, the implicit determination of the V-T regression in
EDK does not allow removal of data outliers, which may affect the quality of the regression.
57
Chapter 2
Fig. 2.5.5
Figure 2.5.5 illustrates an application of EDK for seismic velocity mapping from an interpreted travel
time map (top left) in a North Sea Field. The example compares a velocity map obtained by least squares
regression of well velocity on travel time (top middle) with the result of applying EDK with an implicitly
calculated linear V T relationship (top right). The interval velocity versus two-way-time regression was
calibrated using velocity data at 13 wells. The velocity difference map shown at the bottom highlights the
residual velocity field implicitly calculated by EDK to honour the well data exactly. In this case, EDK was
applied with a unique neighbourhood configuration. The regression line implicitly calculated by EDK
is therefore consistent with the explicit well-derived regression line shown in blue in the cross-plot. This
consistency is confirmed when comparing the blue regression line in the cross-plot with the red points,
which represent the EDK results at all grid nodes. EDK is a powerful technique but in practice, the automatic determination of the V-T regression is not desirable. Instead we recommend using Kriging with
LVM or cokriging to achieve similar results with the added advantage of direct control on the calibration
between V and T and the possibility to manually remove regression outliers, if required.
2.6 Cokriging
Cokriging is a multivariate extension of kriging also developed by Matheron in the 1960s. The goal
is to predict an unknown rock property at one location by combining measurements of the same
property at other locations with measurements of another variable, which is assumed to be cross-correlated with the first one. The predicted variable (e.g., porosity) is often called the primary variable
and the other variable, which is used to improve the estimation of the primary variable, is called the
secondary variable. Cokriging is particularly suitable when the primary variable is significantly undersampled compared to the secondary one. This explains the popularity of the technique in the context
58
Fig. 2.6.1
of seismic-guided earth modelling, where dense and regularly sampled geophysical attribute data are
combined with sparsely sampled well measurements of the primary rock property of interest (Doyen,
1988). To simplify notation, we will restrict ourselves to situations where we have a single guiding seismic attribute but cokriging can easily be extended to multiple secondary variables.
In its most general form (Figure 2.6.1), cokriging involves a generalized linear regression mixing n primary data (well) and m secondary (seismic) data values. As with kriging, the regression
weights are determined by solving a set of normal equations resulting from the minimization of the
mean square prediction error. The books by Journel and Huijbregts (1978) and Chils and Delfiner
(1999) provide full details on these equations. In practice, the general form of cokriging is rarely
used because it requires the extraction and modelling of three spatial covariance functions: the
auto-covariance of the primary variable, Cxx, that of the secondary variable, Czz, and the cross-covariance Cxz between primary and secondary variables (Czx is usually assumed to be equal to Cxz). In 1992,
Xu et al., proposed a useful simplification called Collocated Cokriging (CCK) (Figure 2.6.2) where
the geo-regression includes only z0, the secondary attribute measured at the estimation location uo.
They make an additional Markov-type screening assumption. It leads to the further simplification
that the cross-covariance function is a scaled version of the auto-covariance of the primary variable,
as shown in Figure 2.6.3. Solution of the simplified normal equations (shown on the same figure)
requires only the knowledge of
1. the auto-covariance of the primary variable,
2. the variance of the secondary variable and
3. the local coefficient of correlation between primary and secondary variables.
59
Chapter 2
Fig. 2.6.2
Collocated Cokriging
Fig. 2.6.3
60
Fig. 2.6.4
A further benefit of the collocated approach is that the simplified normal equations are more stable than the general implementation, where the cokriging system involves closely spaced redundant
secondary data and is therefore often ill-conditioned.
The approach proposed by Xu et al. still requires the solution of an extended set of normal equations compared to kriging. We will now show that a further simplification is possible and we can obtain
the same collocated cokriging estimate by direct Bayesian update of kriging (Doyen et al., 1996). The
linear update only requires the knowledge of the kriging variance and the coefficient of correlation,
xz. No cokriging system needs to be solved and no reference to spatial cross-covariance is required. To
introduce the Bayesian framework, we first revisit Simple Kriging (SK) in a Gaussian context. As shown
in Figure 2.6.4, the kriging estimate and kriging variance define the mean and variance of a local
Gaussian conditional distribution at the estimation location. The spread of the distribution is linked
to the uncertainty in predicting xo from the well data alone. To simplify notation, we have dropped the
subscripts o so that x is equivalent to xo and xk to x0k. We also use the superscript k instead of sk to refer
to simple kriging. Now suppose (Figure 2.6.5) that in addition to the well data, we observe a seismic
attribute, z, at the same location uo. How can we use this extra information to update the PDF constructed from the well data? We assume that the joint distribution f(x,z) is Gaussian and given that we know z,
we can define the range of possible value of x from the conditional distribution f(x|z). We can combine
the two pieces of information using the Bayesian updating rule shown in Figure 2.6.6, where the local
posterior distribution is expressed as a product of a seismic likelihood function and a prior distribution
obtained by kriging the well data. Under our Gaussian assumption, both terms represent Gaussian kernels, but in fact the Bayesian rule is general and can be applied to other types of distributions as we will
see later. The decomposition of the local posterior distribution into a product of local likelihood and local
61
Chapter 2
Fig. 2.6.5
posterior has been proposed by Besag (1986) in the context of statistical image processing. One important
assumption underlies this decomposition: the conditional distribution of z depends only on the local value
x and not on x values at adjacent locations. This is a Markov-type assumption equivalent to the assumption
made by Xu in his simplified implementation of collocated cokriging. In fact, it is easy to show that the
cokriging solution obtained by Bayesian updating is exactly identical to that obtained by Xu.
The Bayesian updating scheme yields a particularly simple formula for the collocated cokriging
estimate xck at one point (Figure 2.6.7): it is obtained as a weighted linear combination of the kriging
estimate xk and the conditional expectation E[x|z]=z, which corresponds to doing a standard linear
regression of x on the local z value. The blending weights are simple functions of the kriging variance
and the coefficient of correlation . In extrapolation situations (i.e., when xk =mx=0 and k2=x2=1), the
cokriging solution reduces to the prediction of classical linear regression whereas cokriging reduces
to kriging (i.e., xck = xk) when =0, as required. This suggests a particularly simple and fast computer
implementation of collocated cokriging:
1. Calculate the kriging estimate xk and kriging variance k2 over the whole 2-D or 3-D
interpolation grid.
2. Update the kriging estimate point-by-point by linear blending with the collocated secondary
attribute z, using the kriging variance and to define the relative weights.
One of the main benefits of this approach is that it decouples the influence of the primary and secondary data and allows fast, interactive sensitivity analysis of the cokriging solution to the coefficient of
correlation between the primary and secondary variable. In Figure 2.6.8, cokriged porosity values from
62
Fig. 2.6.6
Fig. 2.6.7
63
Chapter 2
Fig. 2.6.8
64
Fig. 2.6.9
a 3-D cube are cross-plotted against seismic impedance for three different values of . Another advantage of the approach is that it is easy to incorporate spatial variations in the coefficient of correlation.
Figure 2.6.9 shows porosity kriging and cokriging results in a 3-D model of a Shuaiba reservoir from
Oman (den Boer et al, 1999). The figure also shows corresponding average porosity maps obtained by
averaging a vertical column of cells. The carbonate reservoir is about 55m thick. The reservoir is multilayered with the top layer (shown in the figure) being the most productive. Numerous horizontal producers have been drilled along the top of the reservoir structure. High porosities between 25% and 30%
are observed in the wells. In the top macro-layer, simple kriging with a single constant mean of about
25% was used. This can be seen, along the flanks of the structure (far away from the wells) where the
kriged porosities appear constant. Cokriging was performed using as input a cube of acoustic impedance
obtained by post-stack seismic inversion. The cokriging results (Figures 2.6.9 and 2.6.10) exhibit significantly more lateral porosity variability, resulting from the seismic impedance constraints. On average,
cokriged porosities are also higher than kriged values. Note that both kriged and cokriged 3-D porosity
models tie the well data exactly by construction. Model validation was performed using a well drilled
after the study was completed. Figure 2.6.11 demonstrates the improved porosity prediction achieved
using the seismic information to constrain the 3-D model, as kriging tends to significantly underestimate
the measured porosities. An important point concerning non-stationarity is illustrated in Figure 2.6.12,
using the same carbonate example. The reservoir is multi-layered and shows a decrease of porosity with
depth due in part to diagenetic effects and better porosity preservation in the oil leg. It is therefore
important to perform cokriging layer-by-layer so that the model parameters, such as mean porosity for
kriging and porosity-impedance correlation, are adapted and capture the depth trend. The histograms
of cokriged porosity in each interval clearly show the vertical drop of average porosity with depth.
Fig. 2.6.10
65
Chapter 2
Fig. 2.6.11
66
Fig. 2.6.12
Fig. 2.7.1
67
Chapter 2
Factorial Kriging
Fig. 2.7.2
Fig. 2.7.3
68
3-D, we can use the same filter template at each grid node. The FK weights are therefore the same at
each point and the kriging system is only solved once. In practice, FK is then applied as a simple 2-D
or 3-D filtering operation by convolution or multiplication in the Fourier domain. A key assumption in
the FK equations shown in Figure 2.7.2 is that the spatial covariance of the input image is just the sum
of the covariances of the different components. This is correct only if the different components are
uncorrelated. FK can be generalized to handle spatially cross-correlated components but this extension is rarely applied in practice, as the modelling effort becomes too great.
How do we decide on the decomposition of a geological image into multiple components? This
is an interpretive process based on the analysis of experimental variograms or covariances. In practice, FK application is justified if we can clearly observe a nested covariance structure, corresponding
for example to a short scale and a longer scale of spatial variability, as shown in Figure 1.3.15 or to
geological features with different anisotropy. In the rest of this section, we will show examples of FK
applications for filtering seismic attribute data. The goal is typically to improve the geophysical signal
by removing unwanted structures corresponding for example to acquisition footprint and noise.
Random noise removal is as a special case of FK. In geostatistics, this is sometimes called kriging with measurement errors. A classic example involves the smoothing of seismic velocities derived
from stacking velocity analysis. In this case (Figure 2.7.3), the spatial covariance function is decomposed into the sum of a nugget effect or white noise plus a structure representing the signal. FK is
used to predict the signal and attenuate the random noise. For random noise, the FK equations take
an especially simple form, as all elements of the kriging matrix correspond to the covariance of the
signal, except the diagonal elements, which correspond to the sum of the signal and noise variances.
Fig. 2.7.4
69
Chapter 2
The 1-D synthetic example shown in Figure 2.7.4 illustrates the impact of selecting the correct noise
variance parameter on seismic velocity noise filtering. The true velocity profile and corresponding
noisy data, obtained by adding random noise with variance n2=9, are shown in the top left corner
of the figure. The signal-to-noise ratio is 1.33 and the velocity spatial covariance corresponds to a
spherical model with 3.6 km lateral range. When selecting too low a value of n2 (bottom right), the
FK-filtered velocity profile remains noisy. With too high a value (bottom left), the filtering results are
over-smooth. Selecting the correct value of n2 yields the reconstructed velocity profile shown in the
top right corner of the figure. Kriging with measurement errors can also be used to combine data with
different noise levels: the example shown in Figure 2.7.5 is the same as before except that we now
combine soft data, representing the noisy seismic velocities, with n2=9 and hard data representing accurate velocity measurements at three well locations, with n2=0. The figure depicts the velocity
profile reconstructed by factorial kriging with the pseudo error bounds corresponding to two times the
kriging standard deviation. In this case, the kriging process combines interpolation and noise filtering: the well velocity data are matched exactly and the noise is removed from the seismic velocities.
FK can also be used to remove spatially organized noise. Figure 2.7.6 shows a synthetic example
illustrating the application of FK for acquisition footprint removal. The seismic attribute map (top
right) may be seen as a sum of three components: 1) a smooth component corresponding to the
geological signal with approximately NW-SE anisotropy, 2) N-S stripes representing the seismic
acquisition footprint and 3) random noise. All three components have the same histogram but distinct
spatial covariance structures. An experimental covariance map is calculated and modelled using three
components: 1) an anisotropic model elongated in the N120 direction, corresponding to the geological feature, a thin ellipse stretched in the N-S direction representing the stripes and a nugget term for
Fig. 2.7.5
70
Courtesy of T. Colou
Fig. 2.7.6
Courtesy of T. Colou
Fig. 2.7.7
71
Chapter 2
Courtesy of T. Colou
Fig. 2.7.8
Courtesy of D. Lecerf
72
Fig. 2.7.9
Fig. 2.7.10
the random noise. FK can be used to extract each individual component, as shown in Figure 2.7.7. The
filtered component 1 closely reproduces the geological signal shown at the top.
Figure 2.7.8 depicts an application of FK on a real land 3-D velocity dataset. On the time slice
extracted from the raw dense stacking velocity cube (left), we recognize the presence of both random
noise and organized noise in the form of horizontal and vertical bands related to the acquisition
geometry. A smooth velocity field (right) was recovered by applying FK and removing the random and
spatially organized noise components (middle). Another example where FK is used as a destriping
filter in 3-D is shown in Figure2.7.9. The removal of vertical stripes is clearly visible in the FK filtered
output shown at the bottom.
Jugla et al. (2004) provide an example of geostatistical filtering applied to time-lapse seismic data
from the Alwyn Field (Figure 2.7.10). They apply FK to reduce noise and acquisition artefacts, and
improve 4-D repeatability. They apply FK separately on the 1981 and 1996 datasets and create corresponding noise-filtered seismic cubes. They design the factorial kriging operator for each survey
by performing the variogram decomposition in a time interval not affected by production effects and
then apply the filters over the whole volumes. Comparison of 81-96 energy difference maps, computed
by averaging over a 60 msec window spanning the producing interval, highlights the noise reduction
achieved by FK spatial filtering. Application of FK yields an attribute map more readily interpretable in
terms of 4-D production effects. Magneron et al. (2006) have introduced an extension of FK to account
for variations in structural dips. They propose to use locally calculated dip and azimuth information to
rotate the variogram model components and obtain a dip-steered factorial kriging solution.
73
Chapter 2
74
Fig. 2.8.1
Fig. 2.8.2
Fig. 2.8.3
75
Chapter 2
Fig. 2.8.4
From D. Lecerf
76
Fig. 2.8.5
This scheme provides an automatic decomposition of the covariance functions and an automatic
extraction of the common signal from the two data vintages. AFACK works whether the noise is random or spatially organized, provided that the noises from the two vintages are not cross-correlated.
The factorial cokriging system (Figure 2.8.4) is a standard system of normal equations with the
covariance matrix to be inverted on the left. The cokriging matrix is built from experimental covariances from the 1st survey and 2nd surveys C11 and C22, evaluated for distance vectors corresponding to
the selected filter template and from the cross-covariance between the two surveys, C12. The cokriging
vector on the right-hand side is also constructed from the cross-covariances between all filter operator
points and the estimation location at the centre of the filter template. The spatially invariant filter
weights are obtained by inverting the cokriging matrix once.
Lecerf and Weisser (2003) have proposed an efficient 3-D Fourier domain implementation of factorial cokriging, as illustrated in Figure 2.8.5. They start with two seismic datasets representing different
data vintages over the same area. After Fourier transforming the two input cubes, they compute the
common signal spectrum by applying AFACK to each frequency slice. After processing all frequencies,
they obtain the cube representing the common signal by inverse Fourier transform. Figure 2.8.6 shows
an example from the Statfjord Field (Duffaut et al., 2003) where this technique was applied to merge
data from OBC and streamer surveys covering the same area. The common cube takes advantage of
the signal redundancy between the two acquisitions and shows higher S/N and improved resolution
compared to the two input cubes. It would clearly be a better starting point for interpretation, inversion and reservoir characterization.
Fig. 2.8.6
77
Chapter
3 StochasticSimulation with Seismic Constraints
Simulation with Seismic Constraints
Stochastic
3.1 Overview
Kriging and cokriging are estimation techniques. In reservoir modelling, they are used to calculate best estimates of rock properties by interpolation of well data, with the option of using seismic
attributes to guide the interpolation process in the case of cokriging. The predictions are best in
the sense that they minimize estimation errors in the mean square sense. Geostatistical estimation
techniques act as low-pass or smoothing spatial filters: low values tend to be overestimated while high
values are under-predicted. Adequate representation of extreme values is important when calculating
hydrocarbon in place from an earth model or performing numerical flow simulation. Over the last 15
years, stochastic simulation techniques have become popular as a way of generating numerical models
that better represent subsurface heterogeneities. The basic idea is to generate models that mimic the
spatial variability expected in the reservoir. In stochastic modelling, this is achieved by generating
samples from multi-dimensional random fields whose spatial correlation structure is analogous to that
expected in the reservoir. Stochastic simulation is often said to be conditional in that simulated models
are constrained with well log information, seismic data and other data types such as production profiles or well test data. After conditioning, multiple realizations can be fed into a fluid flow simulator
to evaluate uncertainty in production forecasts. Excellent reviews of stochastic modelling techniques
and subsurface uncertainty quantification workflows include Haldorsen and Damsleth (1990), Omre
(1992), Lia et al. (1997) and Caers (2005).
Our goal here is not to review all the different techniques that have been developed in recent years
for reservoir heterogeneity modelling. Instead, we will focus on some of the most popular techniques for
generating stochastic models constrained with seismic information. Stochastic heterogeneity modelling
is often implemented as a two-stage process (Damsleth et al., 1992). In the first stage, the reservoir facies
architecture is simulated using discrete modelling techniques. In the second stage, simulated discrete
geo-bodies are in-filled with continuous rock properties such as porosity and permeability. Seismic information can be used to constrain both stages of the stochastic modelling process. In this chapter we will
consider stochastic methods for simulating continuous variables from seismic attribute data. In the next
chapter, we will discuss techniques for simulating discrete variables such as litho-facies.
79
Chapter 3
Fig. 3.2.1
Fig. 3.2.2
80
Fig. 3.2.3
the Gaussian PDF defined at each grid node by kriging only the well data. If simulation is done independently at each location, it is clear that the resulting 2-D porosity model will not exhibit any spatial
continuity. The trick in SGS is to have a feedback loop and incorporate previously simulated values
as extra data points for the kriging process. Imagine, as shown in Figure 3.2.1, that we have already
simulated porosity values at three grid nodes (x4 to x6). We select a new pixel at random. The Gaussian
PDF at the new pixel is obtained by kriging from the well data (x1 to x3) and the previously simulated
values along the random path. This feedback loop, which retains previously simulated values as extra
data points, is the key to ensuring that the simulations are spatially correlated. In fact, it can be shown
that simulations generated in this way will approximately reproduce the spatial covariance model used
in the kriging process. By construction, SGS simulations will be conditioned by the well data, i.e., the
simulations reproduce the well observations. Multiple simulations are generated by using different
random paths or by repeated sampling from the local conditional distribution defined at each step.
Figure 3.2.2 shows a 2-D example of stochastic porosity simulation using SGS in a North Sea reservoir. For comparison, the figure also shows the porosity map obtained by kriging the well porosity data.
The kriged map provides a smooth reconstruction of the porosity distribution that underestimates the
expected level of spatial variability in the reservoir, especially away from well control, where simple kriging converges to a constant mean value. In contrast, the simulations attempt to mimic the expected
natural porosity variability; Irrespective of well proximity, they exhibit similar spatial frequency content,
controlled through the choice of the variogram model. While the degree of spatial variability is uniform in a given simulation, the level of variations from one simulation to the next is not. Close to well
locations, multiple simulations look the same. The greatest changes from simulation to simulation are
observed far away from the wells, as expected. In fact, if we construct a histogram from a large number
81
Chapter 3
Fig. 3.2.4
Fig. 3.2.5
82
of simulations at one point, the variance of simulated values will approximately match the well-based
kriging variance. Similarly, if we average a large number of simulations at each grid node, the resulting
mean is approximately equal to the estimate that would be obtained by kriging the well data.
Stochastic simulation generates spatial images of reservoir properties that are visually pleasing. In
particular, they do not show the bull-eyes seen on conventional maps and they exhibit higher frequency details. Furthermore, we get the feeling of visualizing reservoir uncertainty when viewing a sequence
of simulated images. But are these images really representative of reservoir heterogeneities? The spatial
continuity in simulated images depends crucially on our choice of variogram model. Figure 3.2.3 depicts
a series of SGS images constructed using isotropic variograms with different lateral correlation ranges.
The scale of variability is clearly controlled by the variogram choice. In practice, when well data are
sparse, it may be difficult to infer the correct variogram model. We should therefore be cautious when
we speak about modelling spatial uncertainty with stochastic simulation, as we only sample uncertainty
within the limit of an imposed statistical model, which is itself uncertain. In practice, statistical parameters such as the variance and variogram range can also be considered as random variables and allow
variation during stochastic simulation but when well control is very limited, the ranges of simulated outcomes may become extremely wide. A better approach is of course to condition the stochastic simulation
with additional geological information and with seismic data, as we will see next.
Practical Considerations
How do we generate random samples from a univariate Gaussian PDF, as required at each SGS step?
Figure 3.2.4 illustrates the basic concept of generating samples from a random variable specified via
its Cumulative Distribution Function F(x). Using a random number generator, we start by drawing a
random number u from a uniform distribution in the interval [0, 1]. Next, we generate a simulated
value x from x = F-1(u) where F-1 is the inverse of the CDF F(x). This approach can be used for any
univariate PDF, not just for Gaussian variables. When an analytical expression does not exist for the
inverse transform, F-1 is calculated from a discretized empirical CDF.
SGS assumes that we are dealing with Gaussian variables. If this is not the case, a Gaussian transform is applied to the well data and the simulation is conducted from the Normal scores. At the end of
the simulation, the results are back-transformed so that histograms of simulated values approximately
match the histogram of the input well data. This transformation process is illustrated graphically in
Figure 3.2.5: an input well data point x1 is mapped into a standard Normal value y1. At the end of the
simulation process, a simulated value, y2, is transformed back into the original space as x2.
Another practical implementation point for SGS involves the use of a moving search region: for
large grids, it is not possible to keep all previously simulated points in the kriging system at one location. Data and previously simulated values are therefore retained only within a moving search region,
as explained in Chapter 2.
83
Chapter 3
Fig. 3.3.1
Fig. 3.3.2
84
Fig. 3.3.3
The easiest way to incorporate seismic information in SGS is via a 2-D or 3-D trend defined for
example from an inverted acoustic impedance cube: we first apply a porosity-impedance regression
equation (calibrated from well data) cell-by-cell in our 3-D model. Next, we use the resulting seismic
porosity cube to define a Locally Variable Mean (LVM). The LVM is used instead of a constant mean
mx in the SGS procedure and the kriging is performed on the residuals.
An alternative option, often used in practice, is to apply SGS with collocated cokriging (Xu et. al.,
1992). This is also a straightforward extension of SGS, where we use cokriging instead of kriging to
calculate the conditional PDFs at each grid node. In Figure 3.3.2, z represents the seismic attribute
collocated with the simulated grid cell. Compared to kriging-based simulations, the local PDFs will be
narrower, as the cokriging variances are smaller than the corresponding kriging ones. In general the
variability from simulations to simulations will be less for seismic-constrained simulations compared to
simulations conditioned only to well data, reflecting the uncertainty reduction. Compared to the kriging-based SGS porosity simulations shown in Figure 3.2.2, the cokriging-based simulations depicted in
Figure 3.3.3 are indeed far more constrained away from well locations. If we average a large number of
simulations at each grid node, the mean result would be identical to the cokriged porosity map shown
in the same figure. In this case, two seismic attributes were used in constraining the SGS process. The
degree of influence of each attribute on the porosity simulations is directly related to their coefficient
of correlation () with porosity.
85
Chapter 3
86
Fig. 3.4.1
Fig. 3.4.2
87
Chapter 3
Fig. 3.4.3
Fig. 3.4.4
88
complex relationship. Obviously in practice, the technique requires access to sufficient samples from
the joint PDF f(z,x) to work. If well data are too sparse to model the joint PDF, Monte Carlo simulation
techniques can be used to augment the number of data points, as will be explained in Chapter 6.
Seismic Downscaling
Fig. 3.5.1
89
Chapter 3
Fig. 3.5.2
Fig. 3.5.3
90
Volume of Support
Fig. 3.5.4
Column Averages Rescaling
One simple solution proposed by Gorell (1995) assumes that the seismic-derived porosity at one point
represents a vertical average over a column of cells in the 3-D model (Figure 3.5.3). First, Gorell generates a 3-D porosity model from geostatistical interpolation of the log data. Next, he applies a columndependent rescaling factor to all cells in one column to ensure that rescaled porosity values match the
local seismic average. The resulting 3-D model incorporates the fine-scale log data and reproduces the
seismic-derived column-average values. Gorells scheme can be adapted to cells with variable thickness by considering thickness-weighted column-averages instead of arithmetic averages as shown in
the figure. However, the 3-D porosity model will not tie at deviated wells. Furthermore, application of
a constant rescaling factor in each column of cells distorts the histogram of porosity values and may
give rise to vertical banding artefacts in the final model. Finally, any deterministic way of rescaling
individual cell values to reproduce column averages has limited value because there are many possible
fine-scale porosity models that match the vertical average values. A stochastic approach is therefore
preferable to tackle the seismic downscaling problem.
Sequential Gaussian Simulation with Block Kriging
Behrens et al. (1996) have proposed a more rigorous stochastic downscaling method, which accounts
for the volume of support difference between the seismic averages and the fine-scale cell values.
Their simulation method is based on the block kriging technique, originally developed for mining
applications (see for example Journel and Huijbregts, 1978 or Isaaks and Srivastava, 1989).
Figure 3.5.4 introduces the concept of support volume. In a discretized 3-D model, rock properties
like porosity represent average values over each grid block. In practice, we ignore the volume of sup-
91
Chapter 3
Fig. 3.5.5
Fig. 3.5.6
92
Block Kriging
Fig. 3.5.7
port and assume that the cells in the 3-D model represent point values, xi. In general, a seismic-derived
porosity value represents a vertical average over a much larger volume V than the volumes of the individual cells. If we assume that the cells have about the same size and that average cell values can be
considered to be point values, we can express the seismic column average xV as the arithmetic average of
all cell values in the column. The method proposed by Behrens is a generalization of SGS where Block
Kriging is used to account explicitly for the support volume of the seismic average porosities. The basic
concept of Sequential Gaussian Simulation with Block Kriging (SGSBK) is illustrated in Figure 3.5.5.
As in traditional SGS, we visit the cells following a random path. At each step, we simulate a porosity
value x by sampling from a Gaussian conditional PDF, where the conditioning data include original well
data plus previously simulated cell values. This time however, we must incorporate the seismic-derived
column average xV as an extra conditioning constraint. For simplicity, we write the column average as x
and omit the subscript V, representing the volume of a column of cells. As we will see next, we use Block
Kriging to estimate the mean xbk and variance 2bk of the required conditional PDF.
So far, we have used point kriging techniques where spatial covariances represent covariances
between point values (Figure 3.5.6, top left). Block kriging involves correlating random variables with
different support volumes. We need for example to consider the covariance between a random variable representing a column or block average and another random variable representing a point value.
Similarly, we may want to correlate two variables representing averages from two different blocks.
Because we are working with linear volume averages, the block-to-point or block-to-block covariances
are obtained by simple volume averaging of the point covariances, as shown in Figure 3.5.6.
93
Chapter 3
Fig. 3.5.8
Fig. 3.5.9
94
Fig. 3.5.10
Figure 3.5.7 shows the block kriging procedure proposed by Behrens. The block kriging estimate at
one cell is obtained as a weighted linear combination of adjacent cell porosity values and the seismicderived, column-average porosity value of the column containing the cell. The block kriging estimate
and kriging variance are obtained by solving a set of normal equations, calling for point-to-point,
point-to-block and block-to-block covariances. The block kriging estimate and kriging variance are
then used to define the mean and variance of a local PDF, which is used in the traditional sequential
simulation scheme (Figure 3.5.5). By construction, the 3-D simulations approximately reproduce the
column average values extracted from seismic data.
Sequential Simulation with Bayesian Updating
In 1997, Doyen et al. proposed a simplified areal downscaling method using a Bayesian updating
procedure coupled to traditional SGS. Like Behrens, they assume that seismic porosity represents
a vertical average over a column of cells but this time the weights are allowed to vary with depth
(Figure 3.5.8). Furthermore, the technique introduces a noise variance parameter that controls the
degree of reproduction of the column average data. When 2=0, the column average data are reproduced exactly. In practice, exact reproduction of the seismic-derived vertical averages is not desirable as they are not exact measurements. The Bayesian simulation scheme does not require solving
block kriging systems. Instead, simulated values matching approximately column-average data are
obtained by updating point kriging estimates, using a likelihood function associated with the column average data. Simple analytical expressions can be obtained for the mean and variance of the
Gaussian posterior distribution (Figures 3.5.9) at each location. The posteriors are sampled sequentially at each cell location as in traditional SGS. If vertical correlation is ignored in its calculation,
the average seismic likelihood function takes a particularly simple form, as shown in Figure 3.5.10.
95
Chapter 3
Fig. 3.5.11
Fig. 3.5.12
96
Fig. 3.5.13
As a column is filled up with simulated porosity values, the likelihood becomes more and more narrow. When the last cell is simulated in a column, the width of the Gaussian likelihood is determined
from the noise variance.
Figure 3.5.11 illustrates an example of the application of SGS with Bayesian updating for a carbonate reservoir. The map at the top represents the input seismic-derived average porosity values; the
bottom map was obtained by vertical averaging of simulated porosity values in each column. The crossplot of the two sets of map values confirms the good reproduction of the seismic averages by the 3-D
simulation method. Other examples of the application of this technique include the work of Al-Deeb
et al., (2002) and Bahar et al. (2004). Behrens and Tran (1998) have extended the SGS with Bayesian
updating technique to multiple seismic attribute maps, representing for example amplitudes extracted
from the top and bottom of a reservoir interval.
One advantage of the Bayesian updating method is that vertically variable weights can easily be
introduced into the column average data. Figure 3.5.12 shows the impact of selecting vertically variable weights: in this case the weights decrease with depth in the reservoir interval to reflect the fact that
the seismic averages are more sensitive to variations near the reservoir top. The three cross-sectional
porosity simulations were generated from the same vertical average data, shown at the top. In a), the
weights are uniform. On the left-hand side of the section, where the imposed vertical average is high,
we observe a homogeneous distribution of high porosity values across the whole interval. From b)
to c), the high porosity values are increasingly concentrated near the top of the layer, as required by
the averaging weights which drop to zero below a certain depth.
97
Chapter 3
Fig. 3.5.14
98
Fig. 3.5.15
99
Chapter 3
However, implementation requires a rather involved MCMC simulation procedure. The inference of
the interaction coefficients in the MRF model is also problematic. Gilbert and Andrieux (2003) have
proposed a seismic downscaling approach based on spectral decomposition. They perform a temporal
spectral decomposition of the log data into a low-frequency component with bandwidth similar to that
of inverted seismic data and a high-frequency component. They generate 3-D simulations of the lowfrequency components using SGS with cokriging of the low-passed log data combined with inverted
seismic data. They simulate the other component using SGS with kriging of the high-frequency component of the log data. Finally, they combine the independently simulated components. One advantage of their technique is that it accounts directly for the band-limited nature of the seismic data but
the independently simulated high- and low-frequency components will contain information outside
the selected bandwidth, depending on the power spectrum corresponding to their variogram models.
Bishop et al. (2004) have proposed a similar approach implemented in the frequency domain.
Fig. 3.6.1
100
Fig. 3.6.2
FFT-MA Decoupling
Courtesy of D. Psaila
Fig. 3.6.3
101
Chapter 3
is called the FFT-Moving Average method or FTT-MA (Le Ravalec et al., 2000). The basic principle
of the MA method (see for example Oliver, 1995) is described in Figure 3.6.1: a spatially correlated
Gaussian random field is obtained by convolving a Gaussian white noise with a 2-D or 3-D filter operator. The shape (or impulse response) of the filter, g, defines the spatial correlation structure of the
generated random field. More precisely, the auto-correlation of output random field, cyy, is equal to the
convolution of the filter operator with itself. Applying the convolution operator in the spatial domain
is fairly inefficient. Instead, as proposed by Le Ravalec, the convolution is performed in the frequency
domain using a 3-D FFT technique. The FFT-MA method follows the numerical recipe highlighted in
Figure 3.6.2, where upper case letters are used to represent variables in the Fourier domain. In addition to speed, one advantage of this technique is the decoupling between the white noise generation
and calculations involving the spatial covariance function cyy(h). It is for example possible to change
one parameter of the spatial covariance and use the same white noise realization to generate spatially
correlated random fields with different smoothness (Figure 3.6.3) or correlation lengths. In this case,
simulations with different degrees of smoothness were generated by changing the exponent of the
exponential covariance model. As pointed out by Le Ravalec et al. (2000), another advantage of FFTMA is that it is possible to perturb simulations locally by changing the input white noise values over
a limited sub-volume or sub-area of the simulated field. The gradual deformation method, described
in the next section, takes advantage of this property to introduce local perturbations. One drawback
of the technique is that the simulation grids must be oversized to avoid periodicity problems with the
FFT. We will show applications of FFT-MA in Chapter 6, where it is combined with another technique
called probability field simulation.
102
Courtesy of M. Le Ravalec
Fig. 3.7.1
Fig. 3.7.2
103
Chapter 3
The GDM method has obvious applications for visualizing uncertainty in reservoir modelling: it
is much easier to visualize uncertainty in the spatial distribution of a reservoir property by looking at
slowly varying realizations than by seeing a sequence of images representing independent realizations.
Nevertheless, GDM has been mainly developed in a stochastic optimization framework for production
history matching. In this context, an objective function is constructed, measuring the degree of fit
between production data and dynamic predictions obtained from a 3-D reservoir model by numerically solving flow equations. GDM is used to iteratively perturb reservoir models until a good match
with the production data is obtained. Instead of perturbing individual grid-block values, deformation
parameters, t, associated with a continuous chain of Gaussian realizations, are optimized sequentially
to obtain realizations that reproduce the dynamic data while preserving the spatial continuity characteristics corresponding to the spatial covariance model. Discussion of conditioning reservoir models
to dynamic data is outside the scope of the course; applications of GDM and related optimization
methods to this problem are discussed for example in Le Ravalec et al. (2005), Kretz et al. (2004) and
Wen et al. (2002).
104
Chapter
4 Facies Modelling from Seismic Data Data
Facies Modelling from Seismic
4.1 Overview
Reservoir models are often built following a two-step sequential approach. In the first step, litho-facies
are mapped in 3-D to represent the reservoir architecture and its main flow units. In the second step,
petrophysical properties are simulated on a facies-by-facies basis. Correct modelling of geological
facies in the reservoir model is a key step because porosity and permeability variations tend to be
dominated by property contrasts between the different facies. In the last chapter we reviewed some of
the most popular techniques to model continuous properties such as porosity. In this chapter, we focus
on techniques suitable for litho-facies modelling. The main difference is that we represent lithology
using categorical variables, i.e., variables that take only a finite set of values corresponding to the different litho-facies existing in the reservoir. There are two broad classes of geostatistical techniques for
lithology modelling: object-based and pixel-based techniques. We concentrate on pixel-based methods
as they provide a more natural way of incorporating seismic information for lithology prediction and
show the link between Bayesian lithology classification and indicator simulation.
105
Chapter 4
Fig. 4.2.1
shown in Figure 4.2.1, where we denote impedance by the letter z, instead of Ip. In practice, variability
in rock properties implies that we will observe an overlap between the ranges of z values of the two
litho-classes. Intuitively, we understand that uncertainty in predicting lithology from impedance will be
directly linked to the degree of overlap between the two ranges of values. In this context, it is natural
to express the observed variability using a probabilistic framework. We therefore model the experimental data histograms using class-conditional probability distributions, which we denote by f (z | ci).
In the simplest case, the two conditional PDFs could be Gaussian distributions with different mean
and variance. Now if we want to predict lithology from seismic impedance, a simple classification rule
consists in selecting class c1 when f (z | c1) > f (z | c2) and select c2 otherwise. This is called Maximum
Likelihood (ML) classification, because the function f is called the likelihood function when it is seen
as a function of the class variable ci.
In the ML classification, we make an implicit assumption that the occurrence of sand and shale is a
priori the same. In practice, we will often have information about the net-to-gross ratio in the reservoir.
This information is encoded in the form of prior probabilities for the litho-classes. For example, if the
N/G is 30%, we would set the prior probabilities as p(c1) = 0.3 and p(c2) = 0.7. To combine this a priori
information with the seismic impedance information, we use Bayes rule, which states that the a posteriori probability of litho-class ci is obtained by multiplying the prior probability by the corresponding
likelihood function for each class:
f(z|ci)p(ci)
p(ci|z)=
p(z)
106
Classification Error
Fig. 4.2.2
We can then define a simple classification rule as shown in Figure 4.2.1 based on selecting the lithoclass that has the Maximum A Posteriori (MAP) probability In MAP classification, the probability p(z)
can be ignored, as it simply acts as a normalizing constant that ensures that p(c1| z) + p( c2|z) = 1.
An interesting question is the classification error, as illustrated in Figure 4.2.2. It is intuitively obvious that classification errors will be directly linked to the degree of overlap between the class-conditional distributions f (z | ci). It is easy to demonstrate that MAP classification minimizes the probability
of error. This demonstration is given for example in Duda et al. (2001).
We may use multiple attributes to reduce the ambiguity in the classification process. In that case, z
becomes a vector of attributes (sometimes called a feature vector) and the class-conditional distributions become multi-variate PDFs. Figure 4.2.3 illustrates the concept of ML classification in a situation
where we have two litho-classes and two seismic attributes, z1 and z2, representing for example P-wave
and S-wave impedances. We have assumed that the class-conditional distributions, f (z | ci) can be represented using bi-variate Gaussian distributions, with class-dependent mean vectors and covariance
matrices.
Bayesian Classification in Practice
Next, we review the use of Bayesian classification in the context of 3-D seismic lithology prediction.
Figure 4.2.4 illustrates the main steps of the workflow.
Step 1. Seismic inversion is performed to produce cubes of elastic properties such as Ip and Vp /Vs, which
will be used as input attributes for the classifier.
107
Chapter 4
Fig. 4.2.3
Courtesy of T. Crozat
108
Fig. 4.2.4
Fig. 4.2.5
Step 2. A training set for the classifier is constructed from elastic log data over intervals of known
lithology, or from inverted elastic attributes extracted around the wells. Attribute values corresponding
to the litho-classes (e.g., gas-sand, oil-sand and water-sand) are displayed using cross-plots to assess the
separability of the different litho-classes from the selected elastic attributes.
Step 3. The class-conditional attribute PDFs are modelled from the training set points using parametric or non-parametric techniques.
Step 4. Bayesian classification is applied point-by-point in the 3-D seismic data volume by computing
the local posterior probability (or likelihood) of each litho-class based on the measured values of the
seismic attributes.
The output of this process is a series of litho-probability cubes, which can be used for uncertainty analysis.
We will illustrate this conceptual workflow using a published example involving seismic lithology
discrimination in a turbidite reservoir (Coulon et al., 2006). Along the way, we will discuss some important practical considerations when applying Bayesian classification.
Before inverting the seismic data, a typical project starts with a rock physics feasibility study during
which we investigate whether Bayesian classification is suitable. This study is based on well log data
and involves an exploratory data analysis phase. We construct cross plots of different elastic attributes
and assess visually the feasibility of distinguishing different lithologies or fluid types (Figure 4.2.5).
109
Chapter 4
Fig. 4.2.6
110
Fig. 4.2.7
111
Chapter 4
Kernel Function
Fig. 4.2.8
Epanechnikov Kernel
Fig. 4.2.9
112
Fig. 4.2.10
The concept of KDE can be generalized easily to multivariate PDFs. Figure 4.2.10 illustrates the
3-D case, with x, y and z representing different seismic attributes used as inputs for Bayesian classification. In this product kernel implementation, we use the same function K to smooth the data along
different axes but we may use different kernel widths, represented by hx, hy and hz for the three seismic
attributes.
Selection of the appropriate smoothing parameters is far more important than selection of the
shape of K. Figure 4.2.11 illustrates the effect of changing the kernel width on the estimates of a 2-D
PDF. If h is too small (in this case, we use the same h value along the two axes), the PDF is too noisy
(top left). On the other hand, if h is too large, we obtain a blurred image of the 2-D PDF (bottom right).
In practice, the choice of the smoothing parameter depends on the number of data points and the
spread of the distribution. The books by Scott and Silverman discuss this issue in detail.
Figure 4.2.12 shows the class-conditional PDFs estimated using KDE from the well training data
shown in Figure 4.2.6. The contours correspond to iso-probability values. At this stage, we have not
yet introduced the seismic data but we can run the classifier using only the log data. For example,
we can apply it to predict lithology at the wells used in the training set or at blind wells. In this
context, various cross-validation procedures can be applied to tune the parameters of the classifier,
as explained for example in Avseth et al. (2005). We can also upscale the logs to mimic the drop of
vertical resolution expected with seismic inversion. We can then recalculate the PDFs using the upscaled
training set (Figure 4.2.13, middle) and re-assess the performance of the classifier by applying it to the
blocked logs.
113
Chapter 4
Fig. 4.2.11
114
Fig. 4.2.12
Fig. 4.2.13
Before applying the classifier in 3-D, we need to verify that the PDFs are properly calibrated to
account for the limited vertical resolution of inverted seismic attributes compared to the fine-scale log
data. This can be done in several ways. First, we check the quality of inversion results at well locations
using QC displays as shown in Figure 4.2.14. In this case we have used a blocky inversion that works
directly in a stratigraphic grid. The match of the inversion at the wells is good and we can detect the
low Vp /Vs (or low Poisson ratio) indicative of hydrocarbon sands. Another important QC (Figure 4.2.15)
is to extract inversion results at well locations and compare cross-plots of inverted elastic attributes
with cross-plots constructed from the corresponding log data. In this example, the log and inverted
data points are consistent. If this is not the case, we may need to re-calibrate the classifier, for example
by re-constructing the PDFs from inverted data extracted near the well locations, as shown in Figure
4.2.13 (right). Another important QC is to check that the inverted data range in the full 3-D volume
is consistent with the overall range of the modelled PDFs. As shown in Figure 4.2.16, this is achieved
for example by navigating through the inverted volumes, section by section and checking that the
inverted data range falls within the support region of the modelled PDFs for the different litho-classes.
In practice, a low probability cut-off value, , is defined. Inverted data points for which f (z | ci) <
for all classes will not be classified. If a large number of such anomalous points exist in the inverted
attribute volumes, this usually means that the classifier PDFs are not properly calibrated, for example
if the well-based training set is not representative of the natural variability existing in the reservoir.
Figure 4.2.17 shows the results of applying the Bayesian classification technique along a random
traverse intersecting three wells. Input attributes for the classifier are cubes of acoustic impedance
and Poisson ratio obtained from pre-stack inversion. In this case, the classifier output consists of three
litho-probability cubes, corresponding to the a posteriori probability p(ci|z) calculated point-by-point
115
Chapter 4
Inversion QC at Well 1
Fig. 4.2.14
116
Fig. 4.2.15
Fig. 4.2.16
Litho-Probability Cubes
Fig. 4.2.17
117
Chapter 4
from the vector of inverted seismic attribute z = (Ip, Poisson ratio) for the three litho-classes. The classifier also returns the most probable litho-class at each point.
How do we assess the performance of the classifier? We can extract the results of the seismic lithology classification at well locations (Figure 4.2.18) and compare them to the litho-logs (Figure 4.2.19).
Performance of the classifier can be summarized using a classification confusion matrix (Figure 4.2.20).
The diagonal elements represent the classification success rate for each litho-class. In the example,
the oil-sand success rate is 70%. This % is calculated as the proportion of samples correctly classified
at the well locations, using upscaled lithology logs as true reference. The off-diagonal elements give
the mis-classification rates. An element i, j in the matrix corresponds to the % of samples classified as
litho-class cj while the true litho-class is ci. In the example, the success rate for water-sand is only 32%
because 56% of the water-sands at the wells are misclassified as shale. This is not surprising in view of
the overlap between the corresponding attribute PDFs.
Finally, Figure 4.2.21 shows the kind of probabilistic analysis that can be performed from the lithoprobability cubes: the oil-sand probability cube has been thresholded to reveal the areas where the
probability of hydrocarbon occurrence is greater than 80%. This type of analysis can be used for well
planning or hydrocarbon volume calculation.
118
Fig. 4.2.18
Fig. 4.2.19
Fig. 4.2.20
119
Chapter 4
Fig. 4.2.21
120
Fig. 4.3.1
Indicator Kriging
Fig. 4.3.2
121
Chapter 4
Fig. 4.3.3
Fig. 4.3.4
122
we calculate experimental indicator covariances or variograms along different directions and fit them
with a parametric model, as explained in Chapter 1.
Our goal is to predict the probability of the two litho-classes away from wells. We can do this using
the Indicator Kriging (IK) technique, as illustrated in Figure 4.3.2 with an areal example. We obtain
an estimate of the sand conditional probability at pixel i as a weighted linear combination of indicator
data from nearby well control points. As before, the weights are determined from the condition that
the mean square prediction error is minimized. They are calculated by solving a set of normal equations constructed from the indicator spatial covariance model.
SIS (e.g. Journel and Gomez-Hernandez, 1989) is a straightforward generalization of SGS and is
depicted in Figure 4.3.3 using the same 2-D conceptual example. We visit the grid cells sequentially
according to a random path. In each cell i, we calculate a local sand / shale probability distribution by
IK, using the original well data and previously simulated cell values as control points. Next, we draw a
simulated value (1 or 0) at random from this local distribution. The simulated binary value is then used
as an additional control point for the remaining steps. A complete lithology simulation is obtained by
repeating this procedure along the random path until all cells have been visited.
How can we constrain the SIS realizations with seismic attribute data? A first strategy consists in
using seismic attributes to define spatially variable proportion trends for the litho-classes. This is
achieved for example by first deriving a sand probability volume by traditional Bayesian classification
and using it to define a spatially variable sand proportion, sand, for the indicator kriging calculations.
Another approach (Doyen et al., 1994), depicted in Figures 4.3.4 and 4.3.5, directly combines SIS with
the maximum likelihood classification technique presented in the previous section. At each pixel in the
2-D grid, we first calculate sand / shale probabilities by indicator kriging. Next, we calculate the likelihood of each litho-class from the class-conditional distribution of the seismic attribute f (z | shale) and
f (z | sand). We build a posterior PDF by multiplying the kriging-based PDF with the seismic likelihood
and simulate a value at random from this local posterior. This scheme is easy to generalize to multiple
seismic attributes by using multivariate PDFs for f. Furthermore, the PDFs f can be modelled using either
parametric or non-parametric techniques, as explained above. The decoupling between seismic and well
information makes it easy to perform sensitivity analysis. Other modifications of SIS involving cokriging have been proposed to simulate litho-facies with seismic attribute constraints (see for example Fichtl
et al., 1997 or Deutsch, 2002). This requires a transformation of the secondary seismic variable so that
transformed values lie between 0 and 1 and the modelling of cross-covariance functions.
A 2-D example of SIS with Bayesian updating from the Oseberg Field is shown in Figures 4.3.6 to
4.3.9. A seismic amplitude map extracted at the base of the Upper Ness interval was used to constrain
the modelling of channel sands (Figure 4.3.6). Gaussian PDFs, calibrated at 14 well locations, were
constructed to represent sand and shale conditional distributions of seismic amplitude. Figure 4.3.7
shows a first set of seismic-constrained stochastic simulations generated with an exponential covariance model with N-S preferential direction of continuity. When we use a spatially invariant anisotropy
model, we are not able to capture the channel belt sinuosity observed on the seismic amplitude map.
Several authors (see for example Xu, 1995 and the discussion in Deutsch, 2002, p. 178) have therefore
proposed to vary the direction of anisotropy. In practice, this is achieved by adjusting the orientation
of the principal axis of anisotropy and of the kriging elliptical search neighbourhood at each pixel.
In this example, channel orientation was interpreted manually at a number of control points from
the seismic amplitude map. The orientation data were then interpolated to construct a 2-D vector
123
Chapter 4
Fig. 4.3.5
Fig. 4.3.6
124
Channel Sand Simulations - Constrained using Seismic Amplitude and Well Data
Fig. 4.3.7
field, which defines the direction of anisotropy at each pixel during SIS. Figure 4.3.8 shows another
set of lithology simulations using seismic amplitude and the space-varying orientation constraints.
Simulated sand bodies now exhibit more realistic non-linear connectivity patterns, consistent with the
seismic data. To illustrate the impact of the choice of spatial continuity model, a last set of SIS simulations with seismic and orientation constraints is shown in Figure 4.3.9. This time, we have used a
principal correlation range of 2 km instead of 1 km. The simulated channels now exhibit large scale
connectivity patterns compared to the more patchy distribution observed in Figure 4.3.8.
The concept of indicator simulation can be applied to other types of categorical variable than litho-facies.
A 4-D example of SIS with Bayesian updating from the Statfjord Field (Doyen et al., 2000) is presented in
Figure 4.3.10. The goal is to obtain a 3-D image of the progression of a water flood between two dates (97
and 91) corresponding to the acquisition of two 4-D seismic surveys. A water flooding indicator variable
is defined: it takes a value of 1 at grid cells that have been flooded between the two survey times and a
value of 0 otherwise. This variable is assumed known at wells where repeat saturation logs are available or
wells that have been history matched. The two seismic datasets have been inverted to acoustic impedance.
Histograms of 97-91 impedance differences have been created for the swept and unswept classes by
extracting impedance values at well locations. SIS with Bayesian updating combines the well indicator data
and the 4-D seismic impedance data to create multiple realizations of the flooding indicator variable and
compute a water flooding probability volume. Figure 4.3.11 shows the voxels where the 97-91 flooding
probability exceeds 80%, compared to the saturation changes predicted by the flow simulator (shown as a
series of vertical sections). This is a useful way of validating the flow pattern predicted by numerical flow
simulation with 4-D seismic information. In this example, there is good correspondence between the zones
where the simulator predicts large water influx and the voxels associated with a high flooding probability.
125
Chapter 4
Fig. 4.3.8
Fig. 4.3.9
126
4-D Classification
Fig. 4.3.10
Fig. 4.3.11
127
Chapter 4
SIS was presented in the case of two litho-classes. The method can be generalized easily to more
than two classes by defining a binary indicator variable and associated spatial covariance model for
each litho-class. A nested implementation of SIS is also possible, where we first simulate net and
non-net rock classes and then simulate multiple litho-classes within the net fraction. This cascaded
procedure is useful to help preserve association between specific facies.
128
Fig. 4.4.1
Fig. 4.4.2
129
Chapter 4
Fig. 4.4.3
Fig. 4.4.4
130
Fig. 4.4.5
By construction, individual realizations will approximately reproduce the seismic-derived average sand
proportion map (Figure 4.4.7).
One obvious limitation of TGS is that we only have one single spatial covariance function that of
the Gaussian field x to control the spatial correlation structure of multiple facies. It is therefore not
possible to impose different anisotropy characteristics on individual facies. The Pluri-Gaussian simulation (PGS) method (e.g., Le Loch and Galli, 1996; Thomas et al., 2005) is an extension of TGS introduced to provide more flexibility on anisotropy modelling and more control on the spatial relationship
between facies. The idea consists in applying a threshold mask to two or more Gaussian random fields
to define the facies simulations. Each Gaussian field imposes its spatial correlation structure to one or
more of the facies according to the defined threshold mask. A simple example is depicted in Figure
4.4.8, involving 2 Gaussian fields with different anisotropy. The threshold mask defines a partition of
the x1-x2 plane. The regions in the partition are constructed using threshold values on the two random
fields. Each region defines the range of x1 and x2 values corresponding to one facies. In the example,
two thresholds, t1 and t2 have been used to define three facies regions. A pair of simulated Gaussian
values (x1, x2) at one point (say the points labelled 1 or 2 in Figure 4.4.8) corresponds to a specific location in the mask from which the simulated facies is determined. In the example, the mask was chosen
so that the green facies can be in contact with the two other facies, which was not possible in the TGS
example. The green facies also possesses a different anisotropy direction determined by the Gaussian
field x1. Definition of the multiple thresholds from facies proportions is more complex than for TGS
(see for example Thomas et al., 2005). In principle, PGS-based simulations can also be conditioned to
seismic data using spatially variable thresholds to model seismic-derived proportion data.
131
Chapter 4
Fig. 4.4.6
132
Fig. 4.4.7
Pluri-Gaussian Simulation
Fig. 4.4.8
133
Chapter 4
Courtesy of S. Strebelle
Fig. 4.6.1
Courtesy of S. Strebelle
134
Fig. 4.6.2
Courtesy of S. Strebelle
Fig. 4.6.3
135
Chapter 4
Courtesy of S. Strebelle
Fig. 4.6.4
the multi-point probabilities cannot be inferred from sparse well data. Instead, the required probabilities are obtained by scanning a training image with the data template and counting the number of
occurrences where matching events are observed.
This suggests a straightforward MPS sequential simulation procedure consisting of the following
steps (Figure 4.6.4):
1. Select a pixel location u to simulate at random.
2. Construct the data template from data and previously simulated points falling within the moving
search neighbourhood.
3. Scan the training image to find matching replicates of the data templates and calculate the conditional probabilities p(x(u)=1 | dn) and p(x(u)=0 | dn) by counting the number of corresponding
events.
4. Draw a binary value at random from the local conditional PDF.
5. Add the simulated value as an extra data point.
6. Repeat steps 1) to 5) until all grid points have been simulated.
The scheme is conceptually simple and does not require any variogram modelling or solving of
kriging systems. Conditioning to well data is also straightforward, similar to traditional SIS. Practical
implementation issues such as selection of appropriate geometric templates, use of search trees to
store pre-computed MPS and multi-grid considerations are discussed for example in Strebelle and
Journel (2001).
136
Courtesy of S. Strebelle
Fig. 4.6.5
Courtesy of S. Strebelle
Fig. 4.6.6
137
Chapter 4
Figure 4.6.5 depicts an application of MPS with seismic constraints in the context of a turbidite
reservoir, as presented by Strebelle et al. (2003) and Caers et al. (2003). A training image (left), was
constructed using an object-based simulation method. The image is not conditioned to well information; it simply represents the prior geological information in terms of expected orientation and size of
the turbidite channel belts. In addition, the MPS simulation is constrained to three wells, using channel orientation information interpreted from seismic data and a seismic-derived sand proportion cube.
One MPS simulation is shown and compared to an SIS result using the same well and seismic constraints. The wormy aspect of the MPS is more consistent with the training image and better captures
the expected meandering of the channel belts. Both results honour the variable, seismic-derived sand
proportions, with higher concentrations of sand bodies in zones of high sand probability. Figure 4.6.6
shows a further conditioning step: sand bodies were first interpreted from seismic data using Principal
Component-based cluster analysis. They are used as extra conditioning data points in MPS and help
define the backbone of some of the channel belts. In essence, MPS reconnects and extends elements
of the sand bodies identified from the seismic data, using the multi-point statistics inferred from the
training image to constrain the shape and sinuosity of the sand bodies. This operation could also be
performed with SIS but would be difficult to implement with Boolean techniques. The MPS technique
is attractive but one obvious question is the availability and representativity of the 3-D training image
required to calculate the multi-point statistics.
138
Chapter
5 StochasticInversion
Inversion
Stochastic
5.1 Overview
Geostatistical Inversion (GI) or equivalently stochastic inversion, was introduced in the nineties by
Bortoli, Dubrule and Haas. The basic philosophy of GI is to generate multiple realizations of elastic
properties with high-frequency content that are consistent with both seismic amplitude and well data.
GI techniques are particularly suitable in a reservoir modelling context because they can be directly
constrained with fine-scale log data and deliver impedance results that can be integrated in reservoir
models without downscaling. The availability of multiple elastic models is also useful for uncertainty
analysis. In recent years, the original GI concept has been extended from acoustic to elastic inversion,
with simultaneous inversion of multiple partial angle stacks. Furthermore, the original GI concept,
which was based on a simple Monte Carlo rejection scheme, has been given a more solid theoretical
foundation thanks to the introduction of a Bayesian framework. The Bayesian setting has also led to
more efficient implementations and has made clearer the link between stochastic inversion and deterministic inversion, where a single smooth, band-limited model is derived from seismic data.
Despite its appeal and the recent introduction of commercial software packages, the application
and acceptance of GI in operational contexts remains relatively low. Instead, a single deterministic
inverted model is used to constrain the reservoir model, as explained in previous chapters. In this
chapter, after reviewing the basic GI concept and some of the recent Bayesian inversion developments,
we will discuss how we can exploit multiple GI realizations in stochastic earth modelling workflows and
explore some of their benefits compared to deterministic inversion. We expect that stochastic inversion
will gain in popularity in the next few years, as more commercial software tools become available to
support the required workflows.
139
Chapter 5
Geostatistical Inversion
Fig. 5.2.1
Bandwidth Extension
140
Fig. 5.2.2
Stochastic Inversion
Fig. 5.2.3
1. All realizations tie the well Ip data that are used as control points.
2. The realizations approximately match the spatial continuity model selected for SGS, i.e., the vertical and lateral variogram models.
3. All realizations approximately reproduce the input 3-D seismic amplitude data.
There has been considerable debate in the industry about the value of GI. The technique is usually well accepted by geologists and reservoir engineers, who are interested in exploiting the details
provided by GI results in geomodelling workflows, where they need to model fine-scale reservoir heterogeneities. On the other hand, among geophysicists, there has been some scepticism about the claim
that GI generates high-resolution models with vertical details not generally available from deterministic inversion techniques. The introductory articles by Francis (2006 a, b) discuss this issue, which is
largely based on a misconception about GI. Geostatistical inversion does indeed provide 3-D Ip images
with high temporal frequencies outside the seismic bandwidth, but these high frequencies do not correspond to some super-resolution property of GI. Resolution is obviously limited by the frequency
content of the seismic data. In GI, the temporal high frequencies are simulated in accordance with the
imposed variogram model (Figure 5.2.2). In essence, GI simulates a full-band 3-D Ip model, consistent
with the observed seismic amplitudes; as the high frequency details are uncertain, GI delivers a family of alternative full-band models, all of which are compatible with the seismic data. In fact, if a large
number of GI realizations are averaged together, the resulting, smooth mean model is equivalent to
the result of a conventional, band-limited inversion scheme (Figure 5.2.3).
In GI, high frequencies are not inverted from the seismic data; they are simulated from the spatial
continuity model. In practice, another source of high frequency information in GI comes from direct
141
Chapter 5
incorporation of the well data. In traditional deterministic inversion, wells are only used to define a
low frequency background model. In GI, individual realizations can be conditioned to reproduce finescale log data. Figure 5.2.3 from Lamy et al. (1999) shows the effect of well conditioning on the results
of stochastic inversion. Vertical details observed at the well location are projected away from the well in
accordance with the underlying spatial continuity model. This ability to condition the inversion results
with fine-scale well data is clearly an advantage of stochastic inversion schemes, especially in mature
fields where a large number of wells may be available. In addition to the mean Ip, a pseudo-uncertainty
cube can be created by averaging GI results: for example, the standard deviation of simulated values
can be computed at each point, as shown in Figure 5.2.3. At well locations, the uncertainty will be low
or zero if the simulations are forced to reproduce the log Ip data.
Following the pioneering work of Bortoli, Dubrule and Haas, a number of authors have proposed
modifications to the SGS-based acoustic impedance inversion scheme. For example, Debeye et al.
(1996) and Grijalba et al. (2000) have introduced an iterative, simulated annealing-based method
that works cell-by-cell instead of trace-by-trace. The 3-D grid is visited following a random path during each iteration cycle. At each cell, an impedance value is sampled from a local conditional PDF.
The simulated value is accepted or rejected depending on a temperature parameter and the change
in an objective function that measures the fit between synthetic and real amplitude data. Shtuka and
Mallet (2001) have implemented a trace-by-trace SGS-based approach with improved convergence
by blending multiple simulated impedance traces. Optimal blending weights are determined using a
simple quadratic optimization procedure. However, all these methods, which apply trace-by-trace or
sample-by-sample model perturbations with simulated annealing or simple Monte Carlo acceptance/
rejection, suffer from convergence problems, especially for long inverted time windows and are very
CPU-intensive.
142
Stochastic Inversion
Bayesian Inversion
Courtesy of A. Buland
Fig. 5.3.1
Earth Model
Courtesy of A. Buland
Fig. 5.3.2
143
Chapter 5
Parameter Dependence
Courtesy of A. Buland
Fig. 5.3.3
Spatial Dependence
Courtesy of A. Buland
144
Fig. 5.3.4
Stochastic Inversion
Spatial Dependence
Courtesy of A. Buland
Fig. 5.3.5
Courtesy of A. Buland
Fig. 5.3.6
145
Chapter 5
Courtesy of A. Buland
Fig. 5.3.7
Courtesy of A. Buland
146
Fig. 5.3.8
Stochastic Inversion
Courtesy of A. Buland
Fig. 5.3.9
Fig. 5.3.10
147
Chapter 5
problem can be solved independently for each frequency. The final solution for the posterior mean
and covariance is then obtained by inverse Fourier transform. Details of the technique can be found
in Buland et al. (2003a and 2003b).
The technique can be used to compute a smooth, best estimate model, which corresponds to the a
posteriori mean. Multiple realizations can also be obtained by direct simulation from the posterior distribution in the frequency domain. Figure 5.3.10 shows an example of the application of this Bayesian
inversion scheme to seismic data from the Sleipner Field (Buland et al., 2003b). They simultaneously
inverted 3 partial stacks with angles ranging from 9o to 33o over a 250 msec time window and a 2 msec
sampling interval. They used an exponential covariance with lateral range of 250 m. Temporally, they
used a nested covariance model with effective range of about 9 msec. The figure compares the smooth
posterior mean solution for Vp, Vs and with a solution obtained by simulation from the posterior distribution. Both sets of solutions are equally consistent with the input seismic data but the simulations
exhibit a higher frequency content controlled by the chosen covariance model. Figure 5.3.11 shows the
effect of merging well log information with the inversion results using a Bayesian updating procedure.
In the updated posterior mean (middle), the effect of the well conditioning is evident from the higher
temporal frequency content observed up to a distance corresponding to the lateral correlation length.
On the simulation result (bottom), the frequency content is uniform across the whole section and the
effect of the well conditioning blends in with the simulated high frequency details.
Bulands method is extremely fast because all computations are performed in the Fourier domain.
However, some limitations result from this simplification. First, the seismic data and the model must
both be discretized on the same 3-D grid, regularly sampled in x, y and time. Furthermore, the spatial
148
Fig. 5.3.11
Stochastic Inversion
AVA Approximation
Fig. 5.3.12
Prior Model
Fig. 5.3.13
149
Chapter 5
covariance model for the prior distribution and the noise covariance model must be strictly stationary;
i.e., they are not allowed to vary in space. Williamson et al. (2007) and Escobar et al. (2006) have recently
extended Bulands Bayesian inversion method to work in stratigraphic grids, and allow non-stationary
covariance structures. Like Buland, they rely on a log-linear model (Figure 5.3.12) to relate reflectivity to
elastic properties but use a two-term approximation to the Zoeppritz equation parameterized in terms of
ln(Ip) and ln(Is). They define the prior model using a 3-D Gaussian random field for the log impedance
parameters (Figure 5.3.13). The seismic data to invert consist of a number of partial angle stacks. This
information is integrated in the Bayesian inversion using a Gaussian likelihood function (Figure 5.3.14),
obtained under the assumptions that seismic noise is not correlated from trace to trace or between
angles. They make the further assumption that the noise is uncorrelated from sample to sample in one
seismic trace, resulting in diagonal noise covariance matrices. They also assume that the noise variance
is stationary in time but laterally variable. In practice, noise statistics are therefore summarized using a
S/N map for each input angle stack. If available, well log data can be integrated using a well likelihood
function (Figure 5.3.15). The joint posterior distribution is obtained by taking the product of the seismic
likelihood function, the prior and the well likelihood. Direct calculation of the posterior mean and covariance is not possible due to the size of the matrices to invert, but in this case Fourier transform cannot be
used due to the irregular sampling in time of the model parameters and the non-stationary nature of the
covariance matrices. Instead, a trace-by-trace decomposition of the posterior is applied (Figure 5.3.16),
leading to an SGS-like simulation scheme, where a local posterior PDF is sampled at each trace location to generate multiple realizations of Ip and Is. In essence, this scheme is analogous to the original GI
approach but instead of generating random samples from the prior distribution and rejecting the ones
that do not match the seismic data, the samples are taken from trace-dependent posterior distributions;
this guarantees that the impedance realizations are consistent with the seismic data by construction.
Seismic Likelihood
150
Fig. 5.3.14
Stochastic Inversion
Well Likelihood
Fig. 5.3.15
Fig. 5.3.16
151
Chapter 5
Courtesy of R. Moyen
Fig. 5.3.17
Courtesy of R. Moyen
152
Fig. 5.3.18
Stochastic Inversion
In Williamsons approach, the model framework consists of a stratigraphic grid defined in the time
domain (Figure 5.3.17). Horizontally, the grid is discretized regularly according to the seismic bin size.
In the vertical direction, the thickness of the grid cells is arbitrary but in practice, average cell thickness is between 1 msec and 4 msec. Vertically, the grid may contain multiple intervals, defined using
intermediate horizons. A different cell thickness may be used in each interval together with a different
stratigraphic style such as on-lap, off-lap or proportional variation. The model framework is populated
with initial Ip and Is values by interpolation of low-pass filtered well log data. This background model
defines the mean values of a log-Gaussian random field corresponding to the prior model. Figure
5.3.18 summarizes the workflow for this Bayesian stochastic inversion, which operates in stratigraphic
grids. The main inputs are:
1. A number of partial angle stacks with corresponding angle-dependent wavelets and S/N maps, used
to construct the angle-dependent noise covariance matrices.
2. A smooth 3-D background model for ln Ip and ln Is, with layer-dependent variances, coefficient of
correlations and horizontal and vertical variogram models, fully specifying the log-Gaussian prior
model.
3. Ip and Is log data (with associated uncertainty) that are used to condition the different impedance
realizations.
The Bayesian inversion calculates a posterior distribution by combining the prior model with the
seismic likelihood function corresponding to the input angle stacks and the well likelihood functions.
Multiple Ip and Is realizations are calculated by sampling the posterior distribution. The posterior mean
and standard deviation are then calculated cell-by-cell by averaging simulated impedance values.
Fig. 5.3.19
153
Chapter 5
Figures 5.3.19 to 5.3.22 show an example of the application of this Bayesian stochastic inversion for a deep-water turbidite field. The data include three seismic angle stacks (16 and 30 and 40
degrees), 3 interpreted horizons and log data at 3 wells. Zero-phase wavelets were extracted separately for the three angles stacks, with dominant frequency around 25-30Hz. A 3-D stratigraphic grid
framework was constructed over a 5 km x 3.8 km area, with seismic trace spacing of 12.5 m and an
inversion time window of about 260 msec, and individual layer thickness of 2 msec, resulting in a
total of about 16 106 grid cells. Figure 5.3.19 shows the Ip inversion results along a random traverse
intersecting the three wells, together with the prior model obtained by 3-D kriging interpolation
of low-pass-filtered logs. The higher frequency content of the multiple realizations is clearly visible
compared to the smooth inverted mean model obtained by averaging 100 realizations. Figure 5.3.20
shows histograms corresponding to the mean result and to one individual realization. As expected,
the range of simulated values is wider than that of the mean result. This can have a large impact
when computing statistics than depend on the tails of the Ip distribution, as we will see in the next
section. Inversion results for Is are shown in Figure 5.3.21. In this example, the wells were not used
as conditioning data but the inversion still provides a reasonable tie at the well locations. By construction, the simulated impedance models and the mean solution will approximately reproduce the
different input angle stacks. The seismic sections in Figure 5.3.22 show the input amplitude data,
the synthetic data computed from the posterior mean, the synthetic data corresponding to one realization and the corresponding residuals. In this example, the level of residual energy is fairly low,
meaning that the synthetics closely match the real data. In general, the level of match is controlled
via the noise covariance matrix.
154
Fig. 5.3.20
Stochastic Inversion
Fig. 5.3.21
Fig. 5.3.22
155
Chapter 5
156
Fig. 5.4.1
Stochastic Inversion
Fig. 5.4.2
Fig. 5.4.3
157
Chapter 5
simulated volumes. Instead of using polygon-based classification, it is also possible to combine stochastic inversion with Bayesian lithology classification.
An illustration of this workflow is depicted in Figures 5.4.3 to 5.4.6. It uses the stochastic inversion
results shown in Figures 5.3.19 and 5.3.21 as input. In this example, a sand polygon was defined
from a Poisson ratio versus Ip cross-plot (Figure 5.4.3), as it provides a better space to discriminate
lithology. Figure 5.4.3 also shows 2-D and 3-D images of the sand probability cube derived from stochastic inversion combined with the polygon-based lithology classification. The probability volume has
been thresholded to show only voxels with sand probability above 30%. Figure 5.4.4 depicts the sand
volume histogram computed from multiple lithology realizations, together with lithology simulations
corresponding to the P10, P50 and P90 percentiles.
What have we gained by applying this stochastic workflow? In a traditional workflow, we would
calculate our best estimate of sand volume by applying the polygon-based classification to Ip, Is data
obtained using deterministic inversion. The 3-D sand distribution predicted from the posterior mean
impedance models (Figures 5.3.19 and 5.3.21) is depicted at the bottom left in Figure 5.4.5; the corresponding sand volume is shown by the red arrow in the histogram. The deterministic inversion (i.e., in
this case, the mean of the posterior distribution) significantly underestimates the actual sand volume.
This is a well-known side-effect of deterministic inversion, which provides a smooth reconstruction
of real impedance values. Poor reproduction of the extremes induces a systematic bias in volumetric
calculations, as these calculations involve the application of cut-offs to the tails of the impedance histograms (see for example the discussion in Francis, 2006 a, b). The underestimation of sand volume
from smooth, deterministic inversion is also clearly visible in Figure 5.4.6, where we show sand maps
158
Fig. 5.4.4
Stochastic Inversion
Fig. 5.4.5
Fig. 5.4.6
159
Chapter 5
(averaged between layers 70 to 110) from the P10, P50 and P90 realizations together with the sand
map derived from the mean inversion results.
Barens and Biver (2004) have proposed a similar workflow (Figure 5.4.7) and applied it to a clastic
reservoir from the Gulf of Guinea. They perform a pre-stack geostatistical inversion to generate multiple Ip and Is realizations. From interpreted log and geological facies data, they define 3 dominant seismic facies, corresponding to three polygonal regions in an Ip versus Ip /Is cross-plot (Figure 5.4.8). Each
seismic facies region contains points from several geological facies. They evaluate the relative proportion of each geological facies within each region by counting the corresponding number of points (see
pie charts in Figure 5.4.8). They then apply a nested lithology simulation scheme to each joint Ip, Is
realization. First, they use the polygon-based classification to obtain seismic facies realizations (step 2
in Figure 5.4.8). Next, they use a Truncated Gaussian Simulation (TGS) scheme to simulate litho-facies
within each seismic facies (step 3 in Figure 5.4.9). They use the relative litho-facies proportions defined
within each seismic facies to determine the spatially variable thresholds for TGS.
In addition to global volumetric uncertainty analysis, results of stochastic inversion can be used for
well connectivity analysis. For example, we may want to calculate a histogram of connected sand volume from multiple impedance and lithology realizations. A recent example from the Brenda Field in
the North Sea is provided by Hicks and Francis (2007). They perform stochastic inversion on intercept
and gradient stacks and generate multiple realizations of AI and GI impedances. Using impedance
thresholds, they convert the impedance realizations into lithology simulations. Figure 5.4.10 compares
the results of deterministic and stochastic inversion for acoustic impedance along one vertical section,
together with the calculated sand and oil-sand probability values. Next, they introduce a seed point
160
Fig. 5.4.7
Stochastic Inversion
Fig. 5.4.8
Fig. 5.4.9
161
Chapter 5
Fig. 5.4.10
162
Fig. 5.4.11
Stochastic Inversion
for connectivity analysis and compute a connected oil-sand probability cube, as shown in Figure 5.4.11.
For this purpose, the grid cells connected to the seed point are identified for each lithology realization,
using a seeded region growing algorithm. The connected sand probability is then calculated at each
cell from the relative number of connected realizations.
Continuous variables such as porosity or clay content can also be simulated easily from acoustic
impedance or other elastic parameters obtained by stochastic inversion. For example (Figure 5.4.1), we
can apply a simple impedance-porosity transform point by point to each impedance realization. This
transform can be established by linear least squares regression from well data or using a rock-physics
based petro-elastic model. However, the resulting porosity simulations will not honour the well data.
An alternative approach that guarantees a well tie involves applying SGS with cokriging, using each
impedance realizations as secondary information. An example of this approach is given by Rowbotham
et al. (2003).
Fig. 5.5.1
163
Chapter 5
consistent with the well data. They iteratively perturb the initial model cell-by-cell using a simulated
annealing procedure to improve the match between real amplitude data and synthetic data. At each
cell, they first draw a litho-facies value from a PDF determined by indicator kriging of control points.
Next they simulate density and acoustic impedance from the corresponding litho-facies-dependent
PDFs. Finally they accept or reject the simulated impedance value based on a probability acceptance
test and a temperature parameter determined by the cooling schedule. More recently, Contreras et
al. (2005) and Merletti and Torres-Verdn (2006) have introduced a pre-stack stochastic inversion that
simultaneously inverts for lithology, Vp, Vs and density using an iterative Markov-Chain Monte Carlo
(MCMC) procedure. Figure 5.5.1 shows their lithology (sand / shale) and elastic inversion results
obtained by inverting four partial angles stacks with angles ranging from 6 to 46 degrees for the Marco
Polo Field in the GOM. After stochastic inversion, they co-simulate petrophysical properties such as
porosity, permeability and water saturation from litho-facies-dependent, six-dimensional joint PDFs
of elastic and petrophysical properties. Figure 5.5.1 shows the petrophysical properties, co-simulated
in the reservoir sands.
It is debatable whether a sequential or combined impedance-lithology simulation workflow is better.
The sequential workflow is less demanding and gives the user more control on parameters calibration
for the different steps. On the other hand, the more ambitious combined approach guarantees better
consistency between inverted elastic properties and the lithology-dependent distributions of the elastic
parameters. In the next chapter, we will again discuss workflows combining seismic and rock physics
inversion. These combined workflows are becoming more and more popular in the industry.
164
Chapter
6 Statistical
Rock PhysicsRock Physics
Chapter 6
Statistical
6.1 Introduction
Following the pioneering work of Mavko and Mukerji (1998), Avseth (2000), Mukerji et al. (2001a
and 2001b) and Eidsvik et al. (2002) among others, statistical rock physics has recently emerged as
an important new approach for quantitative seismic reservoir characterization. Statistical rock physics
combines deterministic rock physics relations with geostatistics: the rock physics equations are used to
establish the link between reservoir parameters and seismic properties, while statistical techniques are
used to propagate uncertainty in the rock physics transform and describe spatial variability.
The geostatistical techniques presented in previous chapters have several limitations. In particular, they rely on a purely statistical calibration between seismic attributes and rock properties. The
quality and reliability of this calibration depends on the availability of a statistically representative
sampling of the reservoir properties, which is difficult to achieve in areas with limited well control.
Furthermore, geostatistical prediction is often performed one variable at a time, ignoring implicitly the complex inter-relationships with other reservoir properties and their variability. This leads
to inconsistencies when separately estimating multiple rock properties. In statistical rock physics,
deterministic Petro-Elastic Models (PEMs) are first established to build physically consistent, multivariate relationships between elastic properties and reservoir attributes such as lithology, porosity, fluid saturation and pore pressure. They are then combined with a Monte Carlo simulation to
construct probabilistic petro-elastic transforms, which take into account the natural variability of the
rock properties and uncertainty in the parameters of the rock physics equations. Finally, the probabilistic PEM is coupled with stochastic simulation or Bayesian inversion schemes to generate spatially
consistent 3-D rock property models.
It is difficult to provide a catalogue of proven techniques or a comprehensive view of statistical rock
physics applications in 3-D reservoir modelling because the field is evolving rapidly and complete
operational workflows are not yet widely available. Instead, in this chapter we present some of the
key concepts for combining rock physics with Monte Carlo and stochastic inversion techniques. We
introduce the concepts trough selected case study applications. Presentation of rock physics theory is
outside the scope of the course. We give some simple examples of PEMs but mainly refer to generic
PEMs, as our focus is on the combination of PEMs with geostatistics, not on the rock physics equations
themselves. The books by Avseth et al. (2005) and Mavko et al. (1998) provide excellent introductions
to the practical aspects of rock physics modelling.
165
Chapter 6
Courtesy of D. Hartmann
Fig. 6.2.1
Fig. 6.2.2
166
The book by Avseth (2005) gives a detailed discussion on combining rock physics and Monte Carlo
simulations to create augmented training sets for lithology and fluid classification. Case studies involving this workflow include Mukerji et al. (2001a and 2001b) and Avseth et al. (2001).
We illustrate the concept of rock physics-based MCS for a problem involving the seismic detection of
gas hydrate accumulations. The goal is to assess the feasibility of identifying sediments containing gas
hydrate using seismic inversion, combined with Bayesian classification. We also need to evaluate whether
we can seismically differentiate gas hydrates from water-filled sediments and sediments containing free
gas. In the absence of well data, we rely on a PEM to forward model the elastic properties and P-wave
and S-wave velocities for sediments with different porosity, clay content, and concentrations of free gas,
gas hydrates and water. We use a PEM developed by Dvorkin et al. (1999) for unconsolidated marine
sediments and extended by Ecker et al. (2000) for sediments containing gas hydrates. In this model, gas
hydrate solid particles filling the pore space are treated as being part of the sediment frame. Their effect
is to reduce porosity and modify the elastic moduli of the composite solid matrix. Figure 6.2.1 (a) shows
the predictions of the PEM in the form of a cross-plot between Ip and Vp /Vs. There are two curves colour-coded according to hydrate and free gas saturations. One curve (lower left corner of the cross-plot)
corresponds to sediments saturated with a mix of water and free gas while the upper curve represents
sediments containing gas hydrates and brine. For thermodynamic reasons, we assume that gas hydrates
and free gas cannot coexist in the pore space. The two curves represent deterministic predictions of the
rock physics model assuming constant sediment porosity (= 35%) and clay volume fraction (Vc=10%)
and no uncertainty on the PEM parameters such as fluid compressibility and elastic moduli for pure gas
hydrates and clay minerals. Interpretation of this rock physics template is discussed in Dvorkin et al.
(2003). The main point is that at fixed porosity and clay content, an increase in gas hydrate concentration yields an increase in acoustic impedance and a decrease in Vp /Vs.
In reality, we know that sediment porosity and clay content are spatially variable and that the PEM
parameters are uncertain. We can model this uncertainty using a simple system theory analogy, as
illustrated in Figure 6.2.2: we have a number of uncertain inputs representing basic properties of the
sediments such as their porosity, clay content and different saturations for water (Sw), free gas (Sg) and
gas hydrate (Sh). The sediments may be over pressured so we also need to consider the pore pressure
and overburden pressure as extra input variables. Uncertainty in each input variable is represented
by a PDF whose range of values corresponds to the expected natural variations. In practice, simple
Gaussian PDFs (as shown in the figure) or uniform distributions are used. We represent the equations
of the rock physics model by a generic system transfer function f. The PEM function f links the uncertain inputs to multiple outputs variables such as acoustic impedance and Vp /Vs, which correspond to
the elastic properties we wish to predict. Our goal is to calculate PDFs for output variables such as Ip
by propagating the input uncertainties through the rock physics model. In general, parameters of the
PEM, such as fluid or mineral moduli, are also uncertain. Extra input PDFs are therefore incorporated
to represent the uncertain PEM parameters.
In general, the PEM is non-linear and the uncertainty propagation problem cannot be solved
analytically. Monte Carlo Simulation provides a simple way to propagate all uncertainties in the
PEM. The basic idea, as shown in Figure 6.2.2, consists in generating random samples from the
different input distributions. For each set of input values, we calculate the corresponding output of
the system f and create a histogram of simulated values, which approximates the PDF of the uncertain output. Figure 6.2.1 (b) shows the results of applying the MCS strategy to our gas hydrates
problem. In this case, we have considered that porosity and clay volume are the only uncertain
inputs, characterized by the Gaussian PDFs shown at the top of the figure. The cloud of points in
167
Chapter 6
Courtesy of D. Hartmann
Fig. 6.2.3
Courtesy of D. Hartmann
168
Fig. 6.2.4
Fig. 6.2.5
the cross-plot shows the spread of elastic properties resulting from variations in porosity and clay
content. MCS is a useful tool for sensitivity analysis. We may for example simulate each uncertain
input in turn, leaving the other inputs fixed at their mean values to evaluate the impact of each
variable on the output uncertainty. Figure 6.2.3 shows the results of MCS when only porosity is
simulated with two different levels of variance and Vc kept constant. It is interesting to observe
that while all input variables are described using Gaussian distributions, the output variables will
not in general be Gaussian, due to non-linearity of the PEM. We will come back to this issue in
the next section.
It is easy to see how we can use the probabilistic PEM constructed by MCS to define a training set
for gas hydrates classification. In Figure 6.2.4, we have defined three litho-classes corresponding to
sediments containing free gas, sediments with Sw = 100% and sediments containing gas hydrates. The
bi-variate litho-class conditional distributions of Ip and Vp /Vs, obtained by MCS, have been modelled
using the kernel density estimation technique introduced in the previous chapter. Overlap between the
computed distributions shows that it may be difficult to differentiate water-saturated sediments from
gas hydrates at low hydrate concentration but it should be easy to discriminate sediments containing
free gas in view of the corresponding PDF separation. The MCS strategy is useful in a wide variety of
contexts. For example, it can be used to study the impact of different production scenarios on the 4-D
seismic response.
How do we perform Monte Carlo Simulation in practice? In Figure 3.2.4, we have already illustrated the basic concept of generating samples from a random variable using the inverse CDF transform
method. Here we also need to consider the local correlations existing between the different uncertain
169
Chapter 6
Fig. 6.2.6
input variables. In our gas hydrates example, we cannot ignore the negative correlation that exists
between and the clay content, Vc. A common technique to incorporate inter-variable correlations is
to use a sequential simulation procedure. This procedure is illustrated schematically in Figure 6.2.5.
First, we simulate Vc from the PDF p(Vc) using the inverse CDF method. Next, we simulate from the
conditional distribution p( | Vc) using the same method. In principle, this procedure can be applied
to any bi-variate distribution and is easy to generalize when simulating more than two correlated variables. A particularly simple implementation (Figure 6.2.6) is possible for correlated Gaussian variables,
because analytical expressions can be obtained for the mean and variance of the required conditional
distribution. For non-Gaussian distributions, the kernel estimation technique can be applied to model
the joint PDFs, from which univariate conditional distributions can easily be extracted.
170
Fig. 6.3.1
PEM for Seismic Pore Pressure Prediction
To illustrate the methodology (see also Sayers et al., 2006), we use a case study of seismic pore pressure prediction in a highly over pressured area from the North Sea (Sayers et al., 2003; Doyen et al.,
2003). The 30 km x 30 km project area covers several producing fields with over pressured reservoir
sands, associated with different pressure cells with pressure difference as large as 3000 psi. Pore pressure is predicted from a multi-layered seismic velocity model determined by reflection tomography.
Figure 6.3.1 depicts the inverted tomographic velocity model for the main over pressured interval at
a depth of about 5 km. To predict pore pressure from seismic velocity, we start from an expression
linking P-wave velocity, Vp, to effective pressure, Pe=Po-Pp, which is the difference between overburden
pressure, Po, and pore pressure, Pp. We use an extension of Bowers formula, as shown in Figure 6.3.2,
which assumes that the zero-stress velocity, Vo depends linearly on porosity, , and clay content, Vc.
The coefficients in the equation (a1 to a5) have been calibrated from well log and RFT data from 21
wells located in the project area. The cross-plot in Figure 6.3.2 represents the sensitivity of P-wave
velocity to changes in effective pressure. The velocity data points correspond to upscaled sonic logs
measured in the same formation but in wells with different depths and pore pressures. The reservoir
pressure values have been computed from RFT measurements while the overburden pressure values
have been calculated by vertical integration of density logs from the surface to the depths of interest. We can invert Bowers model to obtain an expression for pore pressure as shown in Figure 6.3.3.
This last equation can be applied point-by-point in a 3-D reservoir model, using as input the seismic
velocity data (Figure 6.3.1) and models for porosity and clay content obtained by 3-D kriging of the
log data (Figure 6.3.4). The overburden pressure in the reservoir interval is calculated cell-by-cell in
the 3-D model by vertical integration of a density cube, constructed by kriging interpolation of logs,
as explained in Figure 6.3.5.
171
Chapter 6
Fig. 6.3.2
Fig. 6.3.3
172
Fig. 6.3.4
Fig. 6.3.5
173
Chapter 6
Uncertainty Propagation
Fig. 6.3.6
Fig. 6.3.7
174
175
Chapter 6
Fig. 6.3.8
Fig. 6.3.9
176
It is interesting to compare PFS with the Sequential Gaussian Simulation (SGS) technique discussed
in Chapter 3. In SGS, spatial correlation between simulated values is achieved by incorporating
previously simulated values in the construction of the local PDFs. In probability field simulation, we
first estimate the local PDFs by kriging the well data; spatial correlation between simulated values is
achieved by using a correlated probability field to sample the pre-defined PDFs.
The PFS technique has several advantages compared to SGS:
1. The calculation of the PDFs is decoupled from the simulation step. This means that PDFs computed in different ways, using geostatistics or other techniques, can be combined together in
PFS.
2. Data conditioning is achieved automatically by specifying local PDFs with zero variance; i.e., when
i (u)=0 in the equation shown in Figure 3.6.9, the simulated value xis(u)=x
i(u).
3. Simulation with PFS is very fast for large grids because the required unconditional simulations can
be obtained using efficient methods such as FFT-MA.
Despite its advantages, PFS has some limitations. First, it does not have a very firm theoretical basis.
Second, it can give rise to artefacts such as local extrema in the vicinity of conditioning data points,
as explained for example in Pyrcz and Deutsch (2001). Nevertheless, PFS is widely applied in practice
due to its conceptual simplicity and ability to handle local conditional distributions no matter how
they have been derived.
In addition to spatial correlation, another important aspect of the simulation procedure depicted
in Figure 6.3.7 is that local correlations between the different input variables must be reproduced.
This is achieved by sequential simulation of the correlated attributes. For example, we first simulate
velocity using PFS. Next, we simulate porosity conditional on the collocated simulated Vp value. Finally,
we simulate the clay content variable locally conditional on both previously simulated values of and
Vp. For this purpose, the required conditional means and variances are calculated using the cokriging
Bayesian updating rule discussed in Chapter 2. The PFS scheme is then applied from the updated
PDFs.
Figure 6.3.10 (a) depicts areal views of a set of stochastic simulations of pore pressure, obtained
using PFS and the workflow in Figure 6.3.7 to propagate all input uncertainties in the rock physics
model. Multiple simulations can be combined to calculate local statistics, such as the probability that
the pore pressure exceeds some critical value above which drilling a well becomes hazardous, as shown
in Figure 6.3.10 (b).
We give another illustration of PFS in the context of this seismic pore pressure case study. In Figure
6.3.5, we estimated the overburden pressure at each depth z by vertical integration of a density cube
obtained by 3-D kriging. How did we obtain an estimate of the uncertainty on the calculated overburden stress? This uncertainty was also needed in the pore pressure uncertainty calculation. It is possible to obtain an analytical expression for the mean square estimation error for Po, as shown in Figure
6.3.11. The first term in this equation represents a thickness-weighted sum of the individual density
kriging errors, which is easy to calculate. However, a second term involving the correlations between
kriging errors at different depths must also be included (e.g., Journel and Huijbregts, 1978, p. 412).
This term is difficult to calculate in practice but cannot be ignored, as kriging errors in adjacent cells
are highly correlated. Instead, overburden uncertainty can be computed using the PFS strategy illustrated in Figure 6.3.12:
177
Chapter 6
Stochastic Pp Simulations
Fig. 6.3.10
Fig. 6.3.11
178
Fig. 6.3.12
1. A Gaussian PDF for density is defined at each point from the kriging estimate and kriging variance
(left).
2. PFS is applied to generate multiple 3-D density simulations by sampling the local PDFs.
3. Each density simulation is converted to a realization of overburden pressure by vertical integration
in each column of cells.
4. The overburden pressure uncertainty is computed from the variance of simulated values at each
point.
5. The error cube is resampled into the reservoir stratigraphic grid for integration with the other
attributes and calculation of pore pressure uncertainty.
This procedure highlights the flexibility of the PFS technique, which can be used to propagate input
uncertainties into different rock physics or geomechanical calculations. In this case, the uncertainty,
P , calculated by PFS is on average two times greater than the value obtained by summing the density
kriging errors with the formula shown in Figure 6.3.11. This shows that in general, we cannot ignore
vertical correlations between the kriging errors. These error correlations are mainly controlled by the
range of the vertical variogram (40 m in this example) used in the 3-D kriging of density.
K
179
Chapter 6
Fig. 6.4.1
Fig. 6.4.2
180
A good general introduction to this technique is given in the book by Morgan and Henrion. (1990).
A useful review of FOM and other related techniques is given by Mishra (1998) in a reservoir engineering context. The technique will again be introduced in the context of our seismic pore pressure
example but it is applicable to any PEM when the predicted variable is expressed as an explicit function of a number of uncertain input variables.
We go back to our system analogy (Figure 6.3.6), where pore pressure Pp is a function of a number
of uncertain input variables xi specified using Gaussian distributions N[x
i (u),i (u)], where u indicates
that the means and standard deviations are spatially variable. We further assume that the joint PDF of
the input vector x at one location is multi-Gaussian with covariances cov[xi,xj]. As mentioned earlier, the
output PDF for Pp at one point will not be Gaussian because the function f linking pressure to velocity
is nonlinear (Figure 6.3.3). However, the output PDF can be approximated by a Gaussian distribution
via linearization (Figure 6.4.1). We consider a Taylor series expansion of f around the mean vector x,
keeping only the first order terms to define Pest as a linear function of the uncertain input variables,
as shown in equation (1) in Figure 6.4.1. It is easy to demonstrate that Pest has a Gaussian distribution
with mean and variance as shown in the figure, where the partial derivatives are evaluated at the mean
vector x. Equation (2) states that the first order estimate of the mean pore pressure is obtained by
evaluating the function f at the mean value of each uncertain input variable. Equation (3) shows that
a linear estimate of the variance of the pore pressure is obtained as a weighted sum of the variances
and covariances of the input variables, with weights representing the sensitivity of the output to the
different uncertain inputs. In general, the covariance terms cannot be ignored for input variables such
as velocity, porosity and clay content, which are significantly correlated. In practice, to quantify pore
pressure uncertainty with the linearized scheme, we just need to compute equation (3) point-by-point
in our 3-D reservoir model, using locally evaluated partial derivatives of the rock physics transform f
(Figure 6.4.2).
It is desirable to assess the relative contribution of each input variable on the output pore pressure uncertainty and the impact of changes in input values on the output uncertainty. The linearized
scheme facilitates such uncertainty and sensitivity analysis. If we ignore the correlations between the
different input variables, we can simplify equation (3) and write it, as shown in Figure 6.4.3, as a sum
of terms representing the uncertainty contribution of each input. In this figure, we have calculated the
uncertainty importance coefficients U of each variable at each point in the 3-D model. The uncertainty
importance volumes have been ranked in terms of decreasing contribution to the global Pp uncertainty
from top to bottom and from left to right.
Uncertainty importance analysis may for example be used to determine which new, improved
measurements would be most significant to more accurately predict pore pressure. In this case, it is
obvious that the overall uncertainty in pore pressure is controlled by uncertainties in seismic velocity, porosity and the coefficient a5, which determines the stress sensitivity of velocity. In comparison,
errors in overburden pressure or clay content have little impact on pore pressure prediction. Another
important point is that the uncertainty importance of each input variable is spatially variable. Spatial
variations of Ui reflect changes in the variance of the corresponding input variable as well as changes
in the partial derivative of the PEM function f with respect to xi. For example, as shown in Figure
6.4.4, zones with fast velocities (e.g., point labelled 1) are associated with higher levels of pore pressure uncertainty than low velocity zones (e.g., point labelled 2) because of the different velocity stress
sensitivity (see cross-plot).
181
Chapter 6
Fig. 6.4.3
Fig. 6.4.4
182
Fig. 6.4.5
183
Chapter 6
Fig. 6.4.6
Fig. 6.4.7
184
approximation E[f(x)]f(x
)=f(E [x
]) is poor in the presence of model nonlinearities. This is sometimes
called the flaw of averages (Savage 2002; see also Mukerji and Mavko, 2005) and refers to the fact that
plugging the mean inputs into the non-linear model f does not yield the mean output. The overestimation of the variance by the linearized scheme is explained by the fact that the tails of the Gaussian
distribution extend well beyond the hydrostatic and overburden pressure limits. These out-of-range
values are automatically rejected in the stochastic simulation, which yields a more narrow range of
simulated outcomes and hence, a smaller predicted standard deviation. Predictions from the linearized method could be improved by calculating the mean and variance of the Gaussian distribution
truncated on the physical pressure limits. However, this correction would not attenuate the bias in the
position of the mode of the Gaussian PDF, compared to the histogram of simulated values.
Figure 6.4.7 summarizes the pros and cons of applying the stochastic and linearized approaches
when propagating uncertainties in a PEM. In general, the stochastic approach is preferred for quantitative uncertainty assessment and to avoid the flaw of averages but the linearized method remains
useful as a first order approximation and for sensitivity analysis.
185
Chapter 6
Fig. 6.5.1
186
Fig. 6.5.2
Fig. 6.5.3
variables can be reproduced by using a sequential procedure. For example, we first compute p(|Ip,Is)
at each point. We then simulate porosity by applying PFS to these conditional PDFs. Next, we calculate
p(Vc|Ip,Is,) at each point from the joint PDF, using the collocated simulated porosity value as extra conditioning data. We then apply PFS to simulate Vc from the updated posterior. We proceed in a similar
manner for water saturation or other predicted variables.
Bachrach (2006) has given an example of this type of empirical Bayesian approach, combining
rock physics and Monte-Carlo simulation to jointly predict porosity and saturation from inverted
seismic data in gas-charged sediments. The inputs for his rock physics inversion are estimates of Ip, Is
and density obtained by pre-stack inversion of seismic data along a 2-D section. He uses a well-calibrated PEM to link porosity and gas saturation to the elastic properties. By Monte Carlo simulation
he builds empirically joint PDFs for the elastic and reservoir properties, starting with uniform priors
for porosity and saturation. From the joint PDFs, he extracts estimates of porosity and water saturation by taking the local Maximum A Posteriori (MAP) solution at each sample point along the seismic
section; i.e., the values of porosity and saturation corresponding to the maximum of the joint local
posterior distribution, p(,Sw|Ip,Is,). At each point, he also calculates the standard deviations on these
estimates from the empirical posteriors. He accounts for the limited inversion accuracy by smoothing
the empirical PDFs established at the well-log scale. Figures 6.5.2 and 6.5.3 display the MAP porosity and saturation estimates along the seismic section together with a comparison of the predictions
with log data at one well. The rock physics inversion is applied only in the sandy intervals, which have
been previously identified using a Bayesian lithology classification scheme. The predictions match the
well values reasonably well but, as expected, uncertainty on saturation prediction remains high, with
standard deviation values in excess of 20%.
187
Chapter 6
Fig. 6.5.4
188
Fig. 6.5.5
Fig. 6.5.6
Saltzer et al. (2005) has also proposed an empirical Bayesian approach for rock physics inversion.
She uses a cascaded seismic and rock physics inversion scheme (Figure 6.5.4) to invert for both porosity
and clay volume from inverted P-wave and S-wave impedances. She uses the Xu and White (1995) model,
combined with a differential effective medium theory to build the PEM linking Ip and Is to and Vc.
She defines an objective function measuring the fit between inverted impedances and impedances predicted from the rock physics model. Starting with an initial guess for and Vc, she minimizes the objective
function using an iterative procedure, which involves repeated calculation of the partial derivatives of the
forward modelled impedances with respect to and Vc. As they cannot be calculated analytically due to
the model complexity, she calculates them numerically from a table of Ip and Is values obtained by forward
modelling from a complete range of possible and Vc values. In practice, she uses a more complicated
objective function than the simple least-squares error function shown Figure 6.5.5 and derives approximate confidence margins on inverted and Vc from posterior covariance statistics in a multi-Gaussian
Bayesian setting. She also incorporates the effect of pore fluids in her rock physics model using Gassmanns
equation. She assumes that the fluid contact positions are known and calculates the fluid-saturated bulk
moduli with the known fluids. Figure 6.5.5 shows an example of inverted porosity and clay volume results
for a deep water reservoir in West Africa, including a comparison of predicted and actual rock properties at a one well. The log data have been low-pass-filtered to the seismic bandwidth for this comparison.
The error sections and error corridors in the log display correspond to one standard deviation, assuming
multi-Gaussian distributions. Predicted errors are larger in the water leg than in the hydrocarbon interval
because the elastic properties of hydrocarbon-filled rocks are more sensitive to changes in and Vc than
brine-saturated rocks. In Saltzers method, choice of the initial model is important, as she uses a local
optimization scheme and does not explore the full range of uncertainty. The inversion scheme is applied
point by point in the 3-D model and does not use any spatial continuity information.
189
Chapter 6
The Bayesian approach described at the beginning of this section is empirical in that a probabilistic representation of the PEM is constructed by Monte Carlo simulation. A more rigorous but more
complex approach relying on Bayesian networks and Markov Chain Monte Carlo Simulation (MCMC)
has been proposed by Eidsvik et al. (2002 and 2004). They use a rock physics model to define the
link between different reservoir and seismic variables. The rock physics equations are represented by a
Bayesian network (Figure 6.5.6) where each node corresponds to one of the reservoir variable. Arrows
in the graph indicate conditional dependences between the different variables. Each arrow or link is
also associated with a forward model equation g and a Gaussian error term, . For example, in their
model, the S-wave velocity is a function g of density and variable Q representing a sand /shale lithology indicator. With a Gaussian error model, the resulting conditional PDF p(Vs|,Q) is also Gaussian
but with non-linear dependence on the conditioning variables. A forward model (an approximation
to Zoeppritz equations in this case) and Gaussian error model are also used to link the reservoir variables (in blue) to the observed seismic data, ds (AVO attributes in this example). This model results in
a Gaussian seismic likelihood function. Finally, a posterior distribution is constructed by combining
the seismic likelihood function with a prior model. In this example, they use a Markov random field
to model the spatial dependence for the two independent parent variables, representing litho-facies
(Q) and fluid (oil/water) indicator (S). All variables in the network are linked either directly or indirectly to these two variables. It is not possible to obtain an analytical expression for the posterior PDF.
Instead realizations are drawn from it using an iterative MCMC procedure. Bayesian networks provide
an elegant framework for uncertainty propagation and rock physics inversion. However, MCMC computational cost and complexity represent a challenge for operational applications in 3-D. In Eidsvik et
al. (2004), 2-D probability maps of facies and fluid type are derived from AVO intercept and gradient
maps. A similar Bayesian network approach has been applied by Veire et al. (2006) to derive pressure and saturation 2-D maps from time-lapse AVO analysis. Other examples of rock physics inversion in a Bayesian framework include the work of Mazzotti and Zamboni (2003) and Gunning and
Glinsky (2007).
190
Pseudo-Well Methodology
Fig. 6.6.1
Pseudo-Well Methodology
Fig. 6.6.2
191
Chapter 6
Pseudo-Well Methodology
Fig. 6.6.3
192
Fig. 6.6.4
Fig. 6.6.5
are extracted at target level and correlated with variations of reservoir properties. Figure 6.6.2 shows an
example from Julien et al. (2002) of the type of AVO sensitivity analysis possible using this pseudo-well
methodology. The target interval contains three sands. They introduce systematic perturbations of average porosity, water saturation, clay content and thickness in each sand interval, resulting in more than
2000 pseudo-well scenarios. Next, they look at the impact of these variations on the AVO response using
cross-plots of far amplitude versus near amplitude responses, as shown in Figure 6.6.3 for the Top Sand.
The cross-plots reveal that porosity changes will strongly affect both near and far amplitudes while saturation variations will mainly impact the far offset amplitudes. It is also clear that several combinations of
porosity, saturation and clay content will have the same AVO response.
The pseudo-well methodology can be extended beyond sensitivity studies to develop quantitative
lithology and fluid prediction workflows. In Section 6.5, it was assumed implicitly that the PEM or
PEM-derived multivariate PDFs established at log scale are directly applicable to inversion results or
applicable after a simple smoothing correction (Bachrach, 2006). Although the inversion process will
partly remove wavelet effects, deterministic inversion results remain band-limited and may provide
biased estimates of log-scale values. El Ouair and Pivot (2002) have proposed a workflow involving the
inversion of pseudo-well synthetics to obtain a better statistical calibration between inverted results and
rock properties (Figure 6.6.4). Their workflow starts by creating representative 1-D pseudo-wells or 2-D
petro-elastic sections representative of different structural, lithology, porosity and fluid content scenarios. As before, forward modelling is applied to create corresponding pre-stack seismic synthetics.
The synthetics are inverted using the same inversion procedure applied to the real data. Next, the
inverted synthetic impedances are used together with the corresponding pseudo-well properties to create attribute PDFs for each target property such as facies (Figure 6.6.5) or porosity. Finally, the PDFs
193
Chapter 6
Fig. 6.6.6
194
Fig. 6.6.7
are evaluated using the real inversion results to derive lithology or porosity probability cubes. Figure
6.6.6 shows one horizon slice through the litho-probability cubes derived using this approach for a
deep offshore reservoir from the Gulf of Guinea. In this example, acoustic impedance was inverted
from a near angle cube (0-6o) and elastic impedance was inverted from a (24-30o) angle stack. In summary, the benefits of using inverted pseudo-wells in this example are twofold: 1) the pseudo-wells provide an extensive training set for statistical petro-elastic calibration and 2) the calibrated PDFs include
the effect of uncertainties due to seismic inversion.
Pseudo-well simulation can also be used as an engine for stochastic inversion. A recent example
from Spikes et al. (2007) is illustrated in Figures 6.6.7 to 6.6.10, using an exploration example from
offshore West Africa. Their approach uses an empirical Bayesian framework to construct posterior distributions of reservoir properties conditioned by seismic amplitude measurements. Following seismic
interpretation and identification of target geobodies in step 1, they construct a field-specific, well-calibrated rock physics model linking petrophysical properties to elastic properties in the 2nd step. In step
3, they construct uniform prior PDFs for the model parameters, which in this case are reservoir thickness, average porosity, clay content and water saturation. They use a three-layer model with constant
properties for shale encasing the reservoir interval. Each input PDF is sampled regularly to generate
a set of pseudo-wells representing all possible combinations of model parameters. The corresponding
elastic properties are calculated using the PEM. This is followed by trace-by-trace synthetic calculation
(in this case near and mid-angle traces were calculated). In step 4, each real trace is compared with all
the synthetic traces contained in the pseudo-well database. Synthetic traces are accepted or rejected
based on the degree of match measured by cross-correlation. In step 5, a posterior distribution of reservoir properties is constructed at each trace location from all the synthetic pseudo-wells that have been
Fig. 6.6.8
195
Chapter 6
Fig. 6.6.9
196
Fig. 6.6.10
accepted in the seismic matching step. In this example (Figure 6.6.9), the posterior PDFs have a fairly
wide range of values, indicating that large uncertainty remains in predicting the reservoir properties,
especially saturation and clay content. In the last step (Figure 6.6.10), the model parameters corresponding to the maximum of the joint posterior PDF are selected at each trace location and displayed
as 2-D attribute maps. In the workflow proposed by Spikes, use of a simple Monte Carlo acceptance
rule resembles the original approach proposed by Haas and Dubrule (1994) for geostatistical inversion. However, one major limitation of the pseudo-well approach is that it ignores spatial continuity:
posterior solutions are obtained independently at each trace location without spatial constraints. This
partly explains the relatively noisy nature of the attribute maps displayed in Figure 6.6.10.
Fig. 6.7.1
197
Chapter 6
Fig. 6.7.2
Fig. 6.7.3
198
1. Inverted elastic attributes must be converted from time to depth for integration with the reservoir
model. Depth conversion is performed as a separate step using a velocity field, which is usually not
coherent with inversion-derived velocities. Proper alignment of inversion results in depth with the
log data is hard to achieve.
2. The issue of downscaling of the seismic inversion results into the earth model framework is often
ignored. For example, in rock-physics inversion, the PEM constructed at the log scale is applied
sample-by-sample to the inverted data without properly accounting for the band-limited nature of
the inversion results.
3. Sequential inversion workflows do not guarantee that the final reservoir model is fully consistent
with the seismic data; i.e., synthetics calculated from the seismic-constrained geomodel will in general not reproduce the input seismic angle stacks.
In this section we will review in some detail a new workflow for Direct Petrophysical Inversion (DPI) proposed by Bornard et al. (2005) and Colou et al. (2005), which attempts to address some of the limitations of
sequential inversion procedures. We will also review some basic issues involving PEM calibration and scale.
The workflow for petrophysical seismic inversion is illustrated in Figure 6.7.2. We start from an initial fine-scale geomodel defined from a 3-D stratigraphic grid in depth (top left). A PEM is applied to
calculate elastic properties in each cell of the geomodel from stored values of porosity, rock type and
saturations (top middle). Angle-dependent reflectivity series are calculated from the elastic properties
at each trace location through the Zoeppritz equation. The reflection coefficient series are then
converted from depth to time using the velocities stored in the geomodel. Angle-dependent 3-D
synthetics are finally generated by wavelet convolution (top right). Perturbations of the properties of
the geomodel are introduced using a simulated annealing algorithm to optimize the degree of match
between the synthetic and the real angle stacks. After convergence, the final geomodel (bottom right)
honours the observed seismic amplitudes, is consistent with the user-specified PEM and integrates
inversion-based velocities that ensure coherence between the depth and time domains.
In essence, the workflow proposed by Bornard and co-workers amounts to updating the properties
stored in a geomodel defined in depth such that after inversion, the model is consistent with the seismic amplitude data. In theory, they can use any pre-existing geological model as a starting point for
the inversion. In practice, one limitation of DPI is that it does not operate directly from the geomodel
framework. Instead, the initial geomodel properties need to be resampled into a stratigraphic grid
with vertical columns of cells regularly distributed according to the seismic CDP spacing.
Scale of Application of the PEM
A key component of DPI is the use of a PEM during the inversion process to link the reservoir properties stored in the geomodel (e.g., porosity, rock type, fluid saturations) to the elastic parameters, as
illustrated in the rock physics template (Odegaard and Avseth, 2004) shown Figure 6.7.3. Grey points
in the Ip vs Vp / Vs cross-plot are log data, with separate clusters representing different rock types, such as
shale and sand with different fluid content. The coloured lines in the cross-plot are the projections of
the PEM equations, calibrated to the well log data for the different rock types. The lines are graduated
in terms of changes in porosity, which in this case is the target variable for the petrophysical inversion.
During inversion, porosity perturbations are introduced via the PEM, which is used in forward modelling for the calculation of the elastic properties. In the Ip vs Vp / Vs space, the perturbations therefore
follow the trajectories defined by the PEM. This helps to guarantee that we obtain a meaningful range
of values in the inversion results.
199
Chapter 6
Fig. 6.7.4
An important point in DPI is that the PEM is established from log data and is applied at the fine
vertical scale of the geomodel. In traditional sequential inversion workflows, the PEM is used in an
inversion step, aimed at converting seismic impedances to rock properties. Application of the logderived PEM to band-limited inversion results can be problematic, as the PEM is only applicable at
the fine scale. This is explained graphically in Figure 6.7.4, where we assume a binary sand-shale mix
of lithologies. At fine scale, sands and shales have distinct values of Vp / Vs. However, a band-limited
inversion will produce intermediate values (shaded polygon), representative of the aggregate lithology
measured at the seismic scale. Applying the fine-scale PEM to the inverted data will yield erroneous
results. Upscaling the PEM is also problematic.
PEM Calibration
How do we construct the PEM? As we have seen earlier, a PEM is basically a set of rock physics equations used to perform forward modelling from reservoir properties to Vp, Vs and density. The PEM
typically combines a series of physical and empirical equations. The parameters of the rock model
depend on the mineral elastic properties, fluid properties, rock texture and depth of burial, and need
to be carefully calibrated using well log and core data. In practice, the parameters of the rock model
are calibrated for each litho-facies and the predictions of the PEM are compared to log values of Vp,
Vs and density, as shown in Figure 6.7.5. This detailed calibration of the PEM is often time-consuming
but it is a crucial step that controls the quality of the petrophysical inversion results. Once the PEM
has been calibrated, 1-D well synthetics are calculated and compared to seismic traces extracted near
the well locations. This step is used to verify the coherency between the seismic and log data (Figure
6.7.6).
200
Fig. 6.7.5
Fig. 6.7.6
201
Chapter 6
Fig. 6.7.7
Fig. 6.7.8
202
Fig. 6.7.9
203
Chapter 6
Courtesy of R. Bornard
Fig. 6.7.10
204
Fig. 6.7.11
PEM Generation
Fig. 6.7.12
The objective function is minimized using an iterative optimization procedure based on simulated
annealing. Grid nodes in the stratigraphic grid are visited following a random path at each iteration
cycle. A modification is proposed and accepted or rejected depending on a temperature parameter
and the change in the objective function. A progressive decrease in the temperature ultimately leads
to convergence to the global minimum of the objective function. Perturbations of the model properties such as porosity are not applied on individual cell values. Instead, groups of cells are perturbed
together at the intermediate vertical scale (V2), corresponding to the resolution of the input seismic
data. The scheme is therefore different from stochastic inversion techniques. No attempt is made to
recover high frequencies outside the seismic bandwidth. Instead high frequencies in the final model
come from the initial model derived from the log data. Uncertainty analysis can be performed by considering different geological scenarios as initial models. After optimization, the different results represent alternative solutions consistent with the seismic data. The PEM can also be changed from one
inversion run to the next, in order to test the impact of rock physics uncertainty on the final results.
Case Study Examples
Figure 6.7.9 shows an example of DPI application for a deep-water deposited sandstone reservoir
from the North Sea (Bornard et al., 2005). Only one exploration well is available, with a 17 m thick
oil column and average porosity on the order of 25%. Six partial stacks with angles ranging from 5 to
32 degrees were inverted simultaneously to derive a 3-D porosity model. The inverted interval covers approximately 350 m and is divided into 141 layers. The figure shows inverted porosity maps for
layers near the top and base of the reservoir. Braided channels with significant lateral and vertical
porosity variations have been delineated with DPI. Figure 6.7.10 shows the evolution of the porosity
distribution in one of the layers during the simulated annealing iterations. Starting from a laterally
205
Chapter 6
PEM Generation
Fig. 6.7.13
206
Fig. 6.7.14
constant porosity distribution, DPI quickly converges to the detailed porosity image shown on the
right, at the bottom of the figure.
Figure 6.7.11 introduces a second example from the Troll West Field (Colou et al., 2006). The
Sognefjord Formation is characterized by an alternation of clean, high porosity and permeability sands
and lower quality micaceous sands. The field has a thick gas cap above a thin, 15 m thick oil column.
The target of DPI was to obtain a seismic-consistent 3-D image of the porosity distribution, starting
from an initial geomodel constructed by log interpolation from more than 20 wells. Five input angle
stacks were used in the inversion. A simple linear porosity-velocity relationship was used for the PEM.
This relation was derived from log data (Figure 6.7.12) after Gassmann-based fluid substitution of the
points in the gas and oil legs to 100% water saturation (Figure 6.7.13). During inversion, Gassmanns
equation is applied again to saturate the rocks with the correct fluids based on the estimated contact
positions. Figures 6.7.14 and 6.7.15 show 2-D and 3-D views respectively, comparing the initial porosity
model and the more detailed image obtained with DPI. Figure 6.7.16 compares the seismic residuals
(i.e., difference between real and synthetic data) calculated from the initial and final porosity models
for one of the angle stacks. As expected, the residuals after DPI inversion have significantly lower energy. This means that the derived 3-D porosity model is more coherent with the input seismic data.
Other DPI Approaches
The approach of Bornard et al. (2005) is attractive but does not offer a complete framework for uncertainty analysis. Probabilistic petrophysical inversion methods, such as the Shell proprietary method
Promise, have also been proposed. Although details have not been published (Leguijt, 2001), Promise
is known to be a Bayesian stochastic inversion method which directly inverts for selected petrophysical
Fig. 6.7.15
207
Chapter 6
Courtesy of F. Allo
Fig. 6.7.16
208
Fig. 6.7.17
properties such as porosity, net-to-gross or reservoir layer thickness. Like the DPI scheme of Bornard
et al., the technique iteratively perturbs an initial geomodel in order to minimize the error between
synthetic and real seismic data. The algorithm produces a number of realizations from which an average model and standard deviation can be calculated. Figure 6.7.17 shows an example of Promise inversion in the Schiehallion Field from Evans et al. (2007). Starting from an initial reservoir model in (a),
they first checked its consistency with new seismic data by computing synthetics depicted in (b). Clear
mis-matches were identified between the initial synthetics and the real data shown in (c). Promise was
used to update the Net-to-Gross (N/G) of the original geomodel. The updated (posterior) N/G distribution in (d) yielded synthetics (e) that better matched the real data. Direct seismic inversion for rock
properties is an active research area. Most published examples involve a step-wise approach where one
variable is inverted at a time (e.g., Kleemeyer et al., 2006) and where full multi-variate uncertainty is
not evaluated.
209
Chapter
7 4-D Earth
Modelling
Chapter 7
4-D Earth Modelling
7.1 Overview
Traditionally, earth models are used to store static reservoir properties such as porosity, lithology and
permeability. Recently, we have seen the emergence of the concept of the 4-D earth model, which
integrates both static and dynamic reservoir properties in the earth model framework. The 4-D earth
model concept has been driven by the need to integrate time-lapse seismic data and reservoir flow
simulation predictions in a shared framework. Figure 7.1.1 illustrates a 4-D earth modelling workflow
in a situation where a base (pre-production) seismic survey and a monitor seismic survey, acquired a
few years after the start of production, are available. An initial geomodel is constructed representing
reservoir conditions prior to the start of production. Properties stored in this initial model include
porosity, clay content and elastic properties such as Vp and Vs, calculated from a PEM using initial values of reservoir pressure and fluid saturations at time T0. Next, elastic properties stored in the model
are updated to represent the saturation and pressure conditions predicted by the flow simulator at the
time T1 of the monitor survey. This is done by applying the PEM cell-by-cell in the 3-D model using the
new values of the dynamic properties. Finally, two sets of 3-D synthetics are generated, corresponding
to the seismic response before and after production started. Synthetic differences represent the effect
of fluid movement and pressure change predicted by the flow simulator. They are compared to real
seismic differences in order to validate the geomodel and flow simulation predictions.
Early examples of this 4-D workflow include the work of Lumley et al. (1994), Gawith and Gutteridge
(1996) and Watts et al. (1996). Although commercial software tools do not yet fully support complete
simulator-to-seismic workflows, a number of proprietary tools and experimental workflows have been
developed in this area. Examples include the work of Biondi (1998), Al-Najjar et al. (1999), Yuh et al.,
(2000), Brechet et al., (2003), Gouveia et al., (2004) and Toinet et al., (2004). In addition to performing a
seismic validation of the flow simulation model, the concept of the 4-D earth model is useful in the context of feasibility studies where the goal is to plan future 4-D seismic surveys and evaluate the optimum
time-lapse intervals between successive surveys, and the expected magnitude and spatial distribution of
the 4-D signal.
Constructing a realistic 4-D earth model is not an easy task. A first challenge is the mapping of data
between different grid representations (Figure 7.1.2) corresponding to the flow simulation, earth model
and seismic grid frameworks. Earth models will typically have several millions of grid cells with cell dimensions of about 100 m x 100 m x 1 m. Due to computational limitations, flow simulation models will typically
be much coarser laterally, with grid block size of about 200 m x 200 m. Finally, seismic data are sampled
on regular 3-D grid with 25 m x 25 m or 12.5 x 12.5 m trace spacing and 4 msec sampling interval along
the time axis. Another difficulty is that the link between dynamic properties predicted by the flow simulator in the elastic properties is not direct. Instead, we need to use a PEM calibrated to core and log data to
transform pressure and saturation values into estimates of elastic properties. Construction of a PEM representative of the local geology and complex production mechanisms is not a straightforward task.
In the next few sections, we focus on the main steps for constructing 4-D earth models, using mainly
illustrations from the Stafjord and Gullfaks Fields. We will not dwell on fundamental rock physics
aspects or 4-D interpretation methodology. Instead, we will highlight some of the specific earth modelling issues that need to be addressed to support effective simulator-to-seismic workflows.
211
Chapter 7
Fig. 7.1.1
212
Fig. 7.1.2
Fig. 7.3.1
213
Chapter 7
Fig. 7.3.2
(i.e., assumed low permeability) are left inactive during the numerical flow simulation. As a result, no
saturation or pressure values are available in the inactive grid blocks and large gaps can exist in the
coarse simulation grid. These gaps need to be in-filled with realistic saturation and pressure values
during the downscaling process, as shown in Figure 7.3.2, which compares the downscaled saturation
field with the original flow simulation model. In this example, downscaling is implemented as a simple
grid-to-grid resampling followed by an interpolation step in the data gaps. More sophisticated downscaling schemes can be applied. Sengupta et al. (2003) find that saturation details at a fine vertical
scale below seismic resolution can affect the seismic response. They also observe that in some cases,
the saturation outputs from flow simulators may be too smooth vertically to construct realistic 4-D synthetics representative of the real seismic response. Instead of the smooth property transfer described
above, they propose to downscale flow simulation saturations to more realistic patchy distributions in
the fine-scale stratigraphic grid. For this purpose, they use an empirical approach where they distribute hydrocarbon saturations (gas in their case) preferentially in high porosity sub-intervals instead of
homogeneously across all cells of the finer model. Castro et al. (2006) and Enchery et al. (2007) have
proposed flow-based downscaling approaches to generate fine-scale saturation and pressure fields
from coarse scale flow simulation grid values.
214
Fig. 7.4.1
PEM depends on a number of parameters such as the bulk moduli and densities of the fluids, the bulk
moduli, shear moduli and densities of the minerals and coefficients defining the stress sensitivity of the rock
frame. In practice, the PEM must be calibrated using log and core data specific to each reservoir.
In general, elastic rock properties are sensitive to changes in both saturation and pore pressure.
The graph in Figure 7.4.2 illustrates the changes we would typically expect as we produce the reservoir.
As water replaces oil in the pore space, we expect to see an increase in P-wave velocity and in acoustic
impedance. The magnitude of the change (typically a few per cent) will depend on several factors
such as the difference in fluid compressibility and the stiffness of the rock frame (e.g., Lumley et al.,
1997). Gassmanns equation is used to calculate the impact of change in fluid saturations on the rock
bulk modulus and seismic velocities. A drop in pore pressure is also accompanied by an increase in
velocity. Unfortunately, there is no universal law to calculate the effect of change in pressure on seismic
velocities. Instead, this part of the PEM often relies on laboratory measurements of ultrasonic velocities on dry rocks as a function of effective pressure, as depicted in Figure 7.4.3. A further complication
arises from the fact that we cannot describe the pressure dependence using a single curve. Instead, if
we measure dry velocities on different cores from the same reservoir in the laboratory, we will often
observe a family of curves, as depicted in Figure 7.4.3. This variability reflects the dependence of the
bulk modulus of the rock frame to changes in porosity, clay content and other parameters. In Figure
7.4.3, we assume that we can describe the velocity pressure dependence using an exponential model
whose asymptote V pdry is expressed as a function of porosity and clay content. In practice, this function
is obtained by empirical regression using a number of core samples with known porosity and clay content. Other empirical models describing the pressure dependence of velocities exist and are described
for example in Eberhart-Phillips et al. (1989) and MacBeth (2004).
215
Chapter 7
Fig. 7.4.2
Fig. 7.4.3
216
Figures 7.4.4 illustrates an example of a rock physics numerical recipe that can be applied cell-bycell in the earth model to calculate P-wave velocity from porosity, clay content, saturation and effective
pressure. This model is of course not universal (e.g., Yuh et al., 2000, Falcone et al., 2004 for other
examples of PEM used in a 4-D context) but is meant to illustrate the types of calculations that are
involved when applying the PEM. In this case, the main steps are as follows:
Calculate V pdry from porosity and Vc using the empirical regression model. In practice, the values of
and Vc define one of the velocity versus pressure curves shown in grey in the figure.
Apply the exponential pressure model (Figure 7.4.3) to calculate the dry rock velocity at the correct
effective pressure Po Pp. In practice the pore pressure Pp comes from the flow simulator whereas
the overburden pressure Po is calculated from a regional depth gradient or from a density cube, as
explained in Chapter 6.
Apply Gassmanns equation to calculate the P-wave velocity at the saturation corresponding to the
value stored in the cell of the geo-model.
Figures 7.4.5 and 7.4.6 show an example of PEM application for calculating changes in P-wave
velocity with water saturation in the Brent interval of the Statfjord Field. Figure 7.4.5 is a vertical
cross-section depicting the evolution of water saturation predicted by the flow simulator from 79 to 97,
after downscaling in the earth model. The three dates correspond to the acquisition times of a base
(pre-production) and two monitor surveys. The water rises from the Western flank to the crest, where
the producers are located. Figure 7.4.6 depicts the corresponding evolution of Vp calculated using a
Stafjord-specific rock physics model (Brevik, 1996). It shows the gradual increase in seismic velocity
that accompanies the water influx at the crest of the field.
Fig. 7.4.4
217
Chapter 7
Fig. 7.4.5
Time-Dependent Vp Predictions
218
Fig. 7.4.6
Impedance-to-Saturation Calibration
Fig. 7.4.7
Fig. 7.4.8
219
Chapter 7
Once the PEM has been applied for the different survey times, we can perform a rock physics sensitivity analysis. For example, we can calculate 3-D grids of saturation and acoustic impedance changes
predicted from the PEM. Figure 7.4.7 shows a fence display of the Upper Brent stratigraphic grid
colour-coded according to 97-91 changes in saturation and acoustic impedance predicted from the
Statfjord 4-D earth model. If the real 4-D data have been inverted, we can also make a direct comparison with impedance values obtained by forward modelling.
Another useful application of the PEM in the context of simulator-to-seismic workflows is the creation of a time-consistent log database (e.g., Brevik et al., 1998), as illustrated in Figure 7.4.8. In mature
fields, a large number of wells may be available. As they are drilled and logged at significantly different times, the measured elastic logs will be affected by production effects. A log database consistent
with the different seismic acquisition times may be created by fluid and pressure substitution, using as
inputs the pressure and saturation values extracted from grid cells intersecting the wells. The synthetic
elastic logs can then be used for example to constrain 4-D inversion work.
Fig. 7.5.1
220
Fig. 7.5.2
Modelling of Overburden
Fig. 7.5.3
221
Chapter 7
resampling the seismic properties into a regular Cartesian grid. Synthetics computed in that way may
not honour the fine-scale heterogeneities represented in the earth model and may contain artefacts
due to resampling problems, for example near faults. A better approach, as illustrated in Figure 7.5.1,
is to perform the synthetic calculation directly from the earth model stratigraphic grids. At each trace
location, all vertical cell intersections are computed explicitly and used to construct the correct reflectivity series in depth. Figure 7.5.2 shows synthetics computed in this way for the Statfjord example,
using a wavelet representative of the real data.
Another important issue for the synthetic calculation is the handling of the overburden. A simple
overburden velocity model may be created by 3-D log interpolation. However, this may not be adequate if we expect significant lateral velocity variations just above the top reservoir. These variations
will affect the top amplitude response and need to be properly captured to obtain realistic 3-D and
4-D synthetics. One possible strategy is to use the results of post-stack seismic inversion to define the
overburden model, as shown in Figure 7.5.3. After depth conversion, inverted acoustic impedances are
mapped into the stratigraphic grid constructed in the overburden. A log-derived velocity-impedance
relationship is then applied to decompose impedance into density and velocity values. Figure 7.5.2
compares real seismic data from the Statfjord Field with synthetics calculated from log-derived and
inversion-derived overburden velocity models. Use of seismic inversion for defining the overburden
velocity field clearly improves the tie between synthetic and real data.
222
Fig. 7.6.1
Fig. 7.6.2
223
Chapter 7
Fig. 7.6.3
224
Fig. 7.6.4
Fig. 7.6.5
In the workflow presented above, the generation of synthetic volumes is tightly coupled to the 4-D
earth model; i.e., synthetics are computed directly from the reservoir model grids. This tight link is
particularly important if we want to quickly compute and compare synthetics corresponding to different geological or flow simulation scenarios. A Gullfaks example is shown in Figure 7.6.5, where two
sets of synthetic amplitude differences are created and compared to real differences to test the validity
of the PEM. In the first set of synthetics, the PEM accounts for both changes of reservoir pressure and
saturation. In the second set, only saturation changes are modelled. Comparison with the real data
shows that the effect of pore pressure decrease is overestimated in the rock physics model; i.e., areas
with pressure drops exhibit large synthetic changes, which do not match the real data. This indicates
that the pressure effect is poorly captured by the PEM response, which has be calibrated to ultrasonic
velocity measurements.
225
Chapter 7
Fig. 7.7.1
The possibility of incorporating 4-D seismic information into history matching as additional dynamic
data is therefore attractive, as it provides images of fluid movements between the wells, albeit at a
vertically averaged scale compared to the layers in the flow simulation model. Huang et al. (1997)
have proposed a combined seismic and production history matching approach where the reservoir
model is updated to match observed production profiles while simultaneously minimizing the difference between real 4-D data and synthetics computed from the flow model. Since then, a number
of approaches have been proposed. They usually involve the minimization of an objective function
composed of two terms: the first term measures the fit between measured and predicted production
data while the second term measures the fit between synthetic and real 4-D seismic data. In the objective function, comparison between predicted and observed 4-D data can be made at different levels,
as illustrated in Figure 7.7.1 from Waggoner (2001). At level 1, real 4-D amplitudes are compared to
synthetic amplitudes, computed from the flow simulator output using a PEM and wavelet convolution. At level 2, we assume that real seismic data have been inverted to acoustic impedance. These
are compared with impedance values forward modelled from the simulator. At level 3, seismic data
are assumed to have been inverted for pressure and saturation (e.g., Landro, 2001), which are directly
compared with the flow simulator outputs. In the last few years, a number of joint production and 4D data inversion methods have been proposed, working at different levels, different scales and using
different optimization strategies. We will not review them here in any detail but give a few references.
Huang et al. (1997) used a simulated annealing optimization procedure with an objective function
measuring the difference between synthetic and real amplitude maps. Landa et al. (1997) assume that
saturation changes can be derived from 4-D data and perform the match at this level. Gosselin et al.
(2000 and 2003) and Dong and Oliver (2003) use gradient-based optimization procedures and work
at the level of inverted seismic impedance. Kretz et al. (2004) use a stochastic optimization approach
226
based on gradual deformation to match 4-D gas saturation indicator data. Stephen et al. (2005) and
Stephen (2006) work with vertically averaged impedance maps and use a Bayesian stochastic framework,
generating an ensemble of models from which inversion uncertainty can be evaluated. Skjervheim et
al. (2005) propose an Ensemble Kalman Filter (EnKF) approach, which allows continuous updating of
the reservoir flow simulation model and uncertainty analysis using an ensemble of models. They use
inverted Poisson ratio maps in the seismic matching process. Based on stream-line flow simulation,
Vasco et al. (2004) have proposed an iterative technique to invert time-lapse amplitude changes for
permeability but they do not integrate well production data into the inversion process. An issue arising
in 4-D seismic history matching workflows is the scale at which the PEM is applied. Some authors (e.g.,
Enchery et al., 2007) apply the PEM in the fine-scale grid of the geomodel, using saturation and pressure fields downscaled from the flow simulation grid. Other authors (e.g., Gosselin et al., 2003) prefer
to apply the PEM directly in the cells of the flow simulation grid. The second approach uses the flow
simulator outputs directly without downscaling but may require upscaling of the PEM, as explained
in Menezes and Gosselin (2006). Joint matching of production and seismic data is clearly an active
research topic and will remain so for the foreseeable future.
227
Future
Directions
Future Directions
Sophisticated software packages for the construction of 3-D earth models have become widely available
in the last few years. We also see better integration of the earth model with seismic interpretation, on
the upstream side, and with flow simulation, on the downstream side. The trend is now for complete
seismic-to-simulator workflows. As a result of these technical advances, it is now much easier to build
earth models than a few years ago. Increasingly, oil companies are moving to earth-model-centric
workflows where the subsurface model is continuously updated throughout the life of the field as new
data become available, and is used to make key field development decisions. The incorporation of seismic information is a crucial element in the construction of earth models. While seismic data are routinely used to define the size and shape of the reservoir, utilization of seismic information to constrain
the modelling of inter-well reservoir heterogeneities is less widespread. Nevertheless, the interest for
constructing seismic-constrained 3-D property models is growing rapidly, due in part to the greater
availability of high resolution 3-D and 4-D seismic data, and powerful pre-stack inversion techniques,
which deliver multiple elastic attribute cubes and allow better lithology and fluid prediction.
In the course we have reviewed some of the most widely used workflows for constraining 3-D earth
models with seismic data. They typically involve a sequential process where seismic data are first
inverted to elastic properties. Inverted data are then depth-converted and transferred into the earth
model framework, where the seismic attributes are used to guide the interpolation of reservoir properties between the wells. Geostatistics provides a number tools for this purpose, including generalized
regression techniques such as cokriging and kriging with external drift, stochastic simulation with
seismic constraints and geostatistical inversion. These techniques are reasonably well-established and
their application can lead to significantly better constrained earth models. So what are the remaining
challenges for seismic data integration in earth models? In particular, what are the main limitations or
bottlenecks with current approaches that need to be addressed in order to make this integration more
widely applicable? In the last few paragraphs, we list some of the future challenges and new directions
in seismic earth modelling.
229
Future Directions
Seismic Downscaling
Although a number of promising seismic downscaling techniques have been recently developed
(Chapter 3), they are not widely used and remain in the hands of R&D specialists. Furthermore,
available techniques usually make simplifying assumptions (e.g., simple linear volume average) about
the seismic averaging process and ignore the physics involved. The issue of constraining a vertically
detailed 3-D earth model with band-limited seismic data therefore remains a major technical challenge. Downscaling is not just a challenge in 3-D but also for 4-D seismic monitoring where the vertically averaged saturation estimates derived from time-lapse data provide limited information about
water fingering in finely stratified reservoirs. We expect this issue to remain at the forefront of R&D
in the coming years.
230
Future Directions
231
References
References
Abrahamsen, P., Omre, H. and Lia, O., 1991, Stochastic models for seismic depth conversion of geological horizons, paper SPE 23138 presented at the Offshore Europe, Aberdeen, UK.
Abrahamsen, P., Hektoen, A.-L., Holden, L. and Munthe, K.L., 1996, Seismic impedance and porosity: support effects: presented at the 5th International Geostatistical Congress of Wollongong,
Australia. E.Y. Boufi and N.A. Schofield (Eds.) Kluwer.
Abrahamsen, P., Hauge, R., Heggland, K. and Mostad, P., 2000, Estimation of gross-rock volume of
filled geological structures with uncertainty measures, SPE Reservoir Eval. & Eng., 3(4), 304-309.
Alabert, F., 1987, The Practice of fast conditional simulations through the LU decomposition of the
covariance matrix, Mathematical Geology 19, 369-387.
Al-Deeb, M., Bloch, G., El-Abd, S., Charfeddine, M., Bahar, A., Ates, H., Soeriawinata T. and Kelkar, M.,
2002, SPE 78510, paper presented at the 10th International Petroleum Exhibition and Conference,
Abu Dhabi, 13-16 October.
Al-Najjar, N.F., Brevik, I., Psaila, E., and Doyen, P.M., 1999, 4D seismic modelling of the Statfjord
Field: initial results, paper SPE 56730 presented at the SPE Annual Technical Conference and
Exhibition, Houston, Texas, 3-6 October.
Almeida, A., 1993, Joint simulation of multiple variables with a Markov-type coregionalization model,
PhD Dissertation, Stanford University, Stanford, CA.
Alsos, T., Eide, A.L., Hegstad, B.K., Najjar, N.F., Astratti, D., Doyen, P. and Psaila, D., 2002, From
qualitative to quantitative 4D seismic aAnalysis of the Gullfaks Field, paper A-28 presented at the
EAGE 64th Conference & Exhibition, Florence, Italy, 27-30 May.
Angerer, E. and Lanfranchi, P., 2003, Fractured reservoir modeling from seismic to simulator: A reality?,
The Leading Edge 22, 684-689.
Armstrong, M., 1998, Basic linear geostatistics, Springer-Verlag, 153p.
Avseth, P., 2000, Combining rock physics and sedimentology for seismic reservoir characterization of
North Sea turbidite systems, PhD Dissertation, Stanford University, USA.
Avseth, P., Mukerji, T., Jorstad, A., Mavko G. and Veggeland, T., 2001, Seismic reservoir mapping from
3-D AVO in a North Sea turbidite system, Geophysics 66, 11571176.
Avseth P., Mukerji, T. and Mavko G., 2005, Quantitative seismic interpretation, Cambridge University
Press, 359p.
Bachrach, R., 2006, Joint estimation of porosity and saturation using stochastic rock-physics modeling,
Geophysics 71, 53-63.
233
References
Bahar, A., Abdel-Aal, O., Ghani, A., Silva, F.P. and Kelkar, M., 2004, Seismic integration for better
modeling of rock type based reservoir characterization: a field case example, paper SPE 88793
presented at the 11th Abu Dhabi international Petroleum Exhibition and Conference, Abu Dhabi,
U.A.E., 10-13 October.
Barens, L. and Biver, P., 2004, Reservoir facies prediction from geostatistical inverted seismic data,
paper SPE 88690 presented at the 11th Abu Dhabi International Petroleum Exhibition and
Conference, Abu Dhabi, U.A.E., 10-13 October.
Behrens, R.A., MacLeod, M.K., and Tran, T.T., 1996, Incorporating seismic attribute maps in 3D reservoir models, paper SPE 36499 presented at the SPE Annual Technical Conference and Exhibition,
Denver, 6-9 October.
Behrens, R.A. and Tran, T.T., 1998, Incorporation seismic data of intermediate vertical resolution into
3D reservoir models, paper SPE 49143 presented at the SPE Annual Technical Conference and
Exhibition, New Orleans, 27-30 September.
Bertrand, C., Tonellot, T. and Fournier, F., 2002, Seismic facies analysis applied to P and S impedances
from pre-stack inversion, SEG Expanded Abstract, 72nd Annual Meeting, Salt Lake City.
Besag, J., 1986, On the statistical analysis of dirty pictures, J. R. Statist. Soc. B, 48, 259-302.
Beucher, H., Fournier, F., Doligez, B. and Rozanski, J., Using 3D seismic-derived information in
lithofacies simulations. A case study, paper SPE 56736 presented at the SPE Annual Technical
Conference and Exhibition held in Houston, Texas, 36 October 1999.
Biondi, B., Mavko, G., Mukerji, T., Rickett, J., Lumley, D., Deutsch, C., Gundes, R. and Thiele, M.,
1998, Reservoir monitoring: a multidisciplinary feasibility study, The Leading Edge 17, 14041414.
Bornard, R., Allo, F., Colou, T., Freudenreich, Y., Caldwell, D.H. and Hamman, J.G., 2005,
Petrophysical seismic inversion to determine more accurate and precise reservoir properties, paper
SPE 94144 presented at the SPE Europec, Madrid, Spain, 13-16 June.
Bortoli, L.J., Alabert, F., Haas, A. and Journel, A.G., 1992, Constraining stochastic images to seismic
data, Proceedings of the International Geostatistics Congress, Troia, Soares A. (ed.), Kluwer publications, Dordrecht, The Netherlands.
Brechet, E., Toinet, S., Ruelland, P. and El Ouair, Y., 2003, 3D Pre-Stack seismic modeling of reservoir grids for 4D feasibility and calibration, paper A-02 presented at the EAGE 65th Conference &
Exhibition, Stavanger, Norway, 2-5 June.
Brevik, I., 1996, Inversion and analysis of Gassman skeleton properties of shaly sandstones using wireline log data from the Norwegian North Sea, SEG Expanded Abstract, BG 2.4.
Brevik, I., Duffaut, K., Furre, A.-K. and Smith, H.R., 1998, Statfjord 4-D establishing time consistency between acoustic log data and 3D seismic, paper 10-16, presented at the 60th EAGE Conference
& Exhibition, Leipzig, Germany, 8-12 June.
234
References
Buland, A. and Omre, H., 2003a, Bayesian linearized AVO inversion: Geophysics 68, 185-198.
Buland, A., Kolbjrnsen, O. and Omre, H., 2003b, Rapid spatially coupled AVO inversion in the
Fourier domain: Geophysics 68, 824-836.
Caers, J., Strebelle, S. and Payrazyan, K., 2003, Stochastic integration of seismic data and geologic
scenarios: A West Africa submarine channel saga, The Leading Edge 22, 192-196.
Caers, J., 2005, Petroleum geostatistics, SPE, 88p.
Calvert, R., 2005, Insights and methods for 4D reservoir monitoring and characterization, SEG/ EAGE
Distinguished instructor short course no 8, 219p.
Castro, S., Caers, J. and Durlofsky, L., 2006, Improved modeling of 4D Seismic response using
flow-based downscaling of coarse grid saturations, paper A024 presented at the 10th European
Conference on the Mathematics of Oil Recovery, Amsterdam, The Netherlands, 4-7 September.
Chambers, R.L., Yarus, J.M. and Hird, K.B., 2000a, Petroleum geostatistics for nongeostatisticians,
Part 1, The Leading Edge 19, 474-479.
Chambers, R.L., Yarus, J.M. and Hird, K.B., 2000b, Petroleum geostatistics for nongeostatisticians,
Part 2, The Leading Edge 19, 592-599.
Chils, J.P. and Delfiner, P., 1999, Geostatistics: modeling spatial uncertainty, Wiley Series in Probability
and Statistics, Wiley & Sons, 695p.
Cole, J., Nebrija, E.L., Saggaf, M.M., Al-Shabeeb, A.N., Den Boer, L. and Doyen, P.M., 2003, Integrated
3D reservoir modeling for Permian Khuff gas development in Ghawar Field, Saudi Arabia, The
Leading Edge 22, 666-669.
Colou, T., 2002, Time-Lapse filtering and improved repeatability with automatic factorial co-kriging (AFACK), paper A-18 presented at the EAGE 64th Conference & Exhibition, Florence, Italy,
27-30 May.
Colou, T., Poupon, M. and Azbel, K., 2003, Unsupervised seismic facies classification: a review and
comparison of techniques and implementation, The Leading Edge 22, 942-953.
Colou, T., Allo, F., Bornard, R., Hamman, J. and Caldwell, D., 2005, Petrophysical seismic inversion,
paper RC 2.7 presented at the SEG/Houston 75th Annual Meeting, Houston, Texas, USA, 6-11
November.
Colou, T., Formento, J.-L., Gram-Jensen, M., Wijngaarden, A.-J., Norenes Haaland, A. and Ona, R.,
2006, Petrophysical seismic inversion applied to the Troll Field, paper presented at the SEG/New
Orleans 76th Annual Meeting, New Orleans, USA, 1-4 October.
Contreras, A., Torres-Verdin, C., Kvien, K., Fasnacht, T. and Chesters, W., 2005, AVA stochastic inversion of pre-stack seismic data and well logs for 3D reservoir modeling, paper T-15 presented at the
EAGE 67th Conference & Exhibition, Madrid, Spain, 13-16 June.
235
References
Coulon, J.P., Lafet, Y., Deschizeaux, B., Doyen, P.M. and Duboz, P., 2006, Stratigraphic elastic inversion for seismic lithology discrimination in a turbiditic reservoir, SEG Expanded Abstract, 76th
Annual Meeting, Houston, Texas, USA, 2092-2096.
Daly, C., Jeulin, D. and Lajaunie, C., 1989, Application of multivariate kriging to the processing
of noisy images, Geostatistics, Volume 2, edited by M. Armstrong , Kluwer Academic Publishers,
Dordrecht, The Netherlands, 749-760.
Damsleth, E., Tjolsen, C.B., Omre, H. and Haldorsen, H.H., 1992, A two-stage stochastic model
applied to a North Sea reservoir, Journal of Petroleum Technology April, 402-408.
Davis, M.W., 1987, Production of conditional simulations via the LU triangular decomposition of the
covariance matrix, Mathematical Geology 19(2), 91-98.
Debeye, H.W.J., Sabbah, E. and Van der Made, P.M., 1996, Stochastic inversion, paper presented at
the 65th Annual International SEG Meeting, Denver, USA, 1212-1215.
De Groot, P.F.M., Bril, A.H., Florist, F.J.T. and Campbell, A.E., 1996, Monte Carlo simulation of wells,
Geophysics 61, 631-638.
Den Boer, L., Doyen, P.M. and Rothenhofer, H., 1999, 3-D seismic porosity modelling using a new
form of cokriging, World Oil May, 77-80.
Deutsch, C. and Journel, A.G., 1992, GSLIB: Geostatistical Software Library and Users Guide, Oxford
University Press, 340p.
Deutsch, C.V., Srinivasan, S. and Mo, Y., 1996, Geostatistical reservoir modeling accounting for precision and scale of seismic data, paper SPE 36497 presented at the SPE Annual Technical Conference
and Exhibition, Denver, USA, October 6-9.
Deutsch, C., 2002, Geostatistical reservoir modelling, Oxford University Press, 376p.
Deutsch, C. V., Tran, T. T. and Pyrcz, M. J., 2002a, Geostatistical assignment of reservoir properties on unstructured grids, paper SPE 77427 presented at the Annual Technical Conference and
Exhibition, San Antonio, Texas, USA, 29 September2 October.
Doligez, B., Beucher, H., Geffroy, F. and Eschard, R., 1999, Integrated reservoir characterization:
improvement in heterogeneous stochastic modeling by integration of additional external constraints, in R. Schatzinger and J. Jordan, eds., Reservoir Characterization-Recent Advances, AAPG
Memoir 71, 333342.
Doligez, B., Beucher, H., Lerat, O. and Souza, O., 2007, Use of a seismic derived constraint: different
steps and joined uncertainties in the construction of a realistic geological model, Oil & Gas Science
and Technology Rev. IFP 62, 237-248.
Dong, Y. and Oliver, D.S., 2003, Quantitative use of 4D seismic data for reservoir description, paper
SPE 84571 presented at the Annual Technical Conference and Exhibition held in Denver, Colorado,
U.S.A., 5-8 October.
236
References
Doyen, P.M., 1988, Porosity from seismic data: a geostatistical approach, Geophysics 53, 1263-1275.
Doyen, P.M., Psaila, D.E. and Strandenes, S., 1994, Bayesian sequential indicator simulation of channel sands from 3D seismic data in the Oseberg Field, Norwegian North Sea, paper SPE 28382 presented at the 69th Annual Technical Conference and Exhibition, New Orleans, LA, USA.
Doyen, P.M., den Boer, L.D. and Pillet, W.R., 1996, Seismic porosity mapping in the Ekofisk Field
using a new form of collocated cokriging, paper SPE 36498 presented at the Annual Technical
Conference and Exhibition, Denver, USA, October 6-9.
Doyen, P.M., Psaila, D.E., den Boer, L.D. and Jans, D., 1997, Reconciling data at seismic and well log
scales in 3-D earth modelling, paper SPE 38698 presented at the Annual Technical Conference and
Exhibition, San Antonio, USA, 5-8 October.
Doyen, P.M., Psaila, D.E., Astratti, D., Kvamme, L.B. and Al Najjar, N.F., 2000, Saturation mapping from 4-D seismic data in the Statfjord Field, paper OTC 12100 presented at the Offshore
Technology Conference, Houston, Texas, USA, 14 May.
Doyen, P.M., Malinverno, A., Prioul, R., Hooyman, P., Noeth, S., den Boer, L., Psaila, D., Sayers, C.M.,
Smit, T.J.H., van Eden, C. and Wervelman, R., 2003, Seismic pore pressure prediction with uncertainty using a probabilistic mechanical earth model, SEG Expanded Abstract.
Dubrule, O., Basire, C., Bombarde, S., Samson, P., Segonds, D. and Wonham, J., 1997, Reservoir geology using 3-D modeling tools, paper SPE 38659 presented at the Annual Technical Conference,
San Antonio TX, USA, 5-8 October.
Dubrule, O., 2003, Geostatistics for seismic data integration in earth models, SEG/EAGE distinguished
instructor short course no 6.
Duda, R., Hart, P. and Stork, D., 2001, Pattern classification, John Wiley & Sons, 654p.
Duffaut, K., Foldal, F., Helgesen, J., Lecerf, D., Oexnevad, G. and Thompson, M., 2003, Processing of
OBC and surface seismic data for reservoir monitoring at the Statfjord field, EAGE paper presented
at the 65th Conference & Exhibition, Stavanger, Norway, 2-5 June.
Duplantier, O., Hadj-Kacem, N. and Vittori, J., 2006, New strategy for seismic facies upscaling to
the reservoir grid scale, paper SPE 101945 presented at the Annual technical Conference and
Exhibition, San Antonio, Texas, USA, 24-27 September.
Dvorkin, J. and Prasad, M., 1999, Elasticity of marine sediments: Rock physics modelling, Geophysical
Research Letters 26, 1781-1784.
Dvorkin, J., Nur, A., Uden, R. and Taner, T., 2003, Rock physics of a gas hydrate reservoir, The
Leading Edge 22, 842-847.
Eberhart-Phillips, D., Han, D.-H. and Zoback, M.D., 1989, Empirical relationships among seismic
velocity, effective pressure, porosity, and clay content in sandstone, Geophysics 54, 82-89.
237
References
Ecker, C., Dvorkin, J. and Nur, A.M., 2000, Estimating the amount of gas hydrate and free gas from
marine seismic data, Geophysics 65, 565-573.
Eidsvik, J., Omre, H., Mukerji, T., Mavko, G. and Avseth, P., 2002, Seismic reservoir prediction using
Bayesian integration of rock physics and Markov random fields: A North Sea example, The Leading
Edge 21, 290-294.
Eidsvik, J., Avseth, P., Omre, H., Mukerji, T. and Mavko G., 2004, Stochastic reservoir characterization
using prestack seismic data, Geophysics 69, 978-993.
El Ouair, Y. and Pivot, F., 2002, Integration of uncertainties due to prestack impedance inversion
in a reservoir characterization process, paper G-23 presented at the EAGE 64th Conference &
Exhibition, Florence, Italy, 27-30 May.
Enchery, G., Le Ravalec-Dupin, M. and Roggero, F., 2007, An improved pressure and saturation down
scaling process for a better integration of 4D seismic data together with production history, paper D003
/ SPE 107088 presented at the 69th EAGE Conference and Exhibition, London, UK, 11-14 June.
Escobar, I., Williamson, P., Cherrett, A., Doyen, P.M., Bornard, R., Moyen, R. and Crozat, T., 2006, Fast
geostatistical stochastic inversion in a stratigraphic grid, 76th SEG Annual International Meeting,
2067-2071.
Evans, A.C., Dankbaar, H. and Stammeijer, J., 2007, 3D and 4D seismic AVO inversion for updating
the Schiehallion reservoir model and prediction of production effects, paper A030 presented at the
EAGE 69th Conference & Exhibition, London, UK, 11-14 June.
Falcone, G., Gosselin, O., Maire, F., Marrauld, J. and Zhakupov, M., 2004, Pertoelastic modeling as
key element of 4D history matching: a field example, paper SPE 90466 presented at the Annual
Conference and Exhibition, Houston, Texas, USA, 27-30 September.
Fichtl, P., Fournier, F. and Royer, J.-J., 1997, Cosimulations of lithofacies and associated reservoir properties using well and seismic data, paper 38680 SPE presented at Annual Technical Conference and
Exhibition, San Antonio, Texas, USA, 5-8 October.
Fournier, F. and Derain, J.-F., 1995, A statistical methodology for deriving reservoir properties for
seismic data, Geophysics 60, 1437-1450.
Fournier, F., Dquirez, P.-Y., Macrides, C.G. and Rademakers, M., 2002, Quantitative lithostratigraphic
interpretation of seismic data for characterization of the Unayzah Formation in central Saudi
Arabia. Geophysics 67, 1372-1381.
Francis, A., 2006a, Understanding stochastic inversion: part 1. First Break 24, 69-77.
Francis, A., 2006b, Understanding stochastic inversion: part 2. First Break 24, 79-84.
Frykman, P. and Deutsch, C.V., 2002, Practical application of geostatistical scaling laws for data integration, Petrophysics 43, 153-171.
238
References
Gancarski, S., Valois, J.P. and Palus, C., 1994, The pseudo-well technique: a tool for statistical calibration of seismic data in a field with limited well control, paper PO55 presented at the 56th Meeting
and Technical Exhibition of the EAEG, 6-10 June.
Gastaldi, C., Roy, D., Doyen, P. and Den Boer, L., 1998, Using Bayesian simulations to predict reservoir
thickness under tuning conditions, The Leading Edge 17, 589-593.
Gawith, D.E. and Gutteridge, P.A., 1996, Seismic validation of reservoir simulation using a shared
earth model, Petroleum Geoscience 2, 97-103.
Gilbert, F. and Joseph, C., 2000, Accounting for seismic scale in stochastic reservoir simulations constrained by seismic data, paper SPE 63290 presented at the Annual Technical Conference and
Exhibition, Dallas, Texas, USA, 1-4 October.
Gomez-Hernandez, J.J. and Journel, A.G., 1993, Joint sequential simulation of multigaussian fields, in
Geostatistics Troia 92, Soares ed., Kluwer publishers, Dordrecht, The Netherlands, 85-94.
Goovaerts, P., 1997, Geostatistics for natural resources evaluation, Oxford University Press.
Gorell, S.B., 1995, Creating 3-D reservoir models using areal geostatistical techniques combined with
vertical well data, paper SPE 29670 presented at the Western Regional Meeting, Bakersfield, CA,
USA, March.
Gosselin, O., Cominelli, A., Van Der Berg, S., and Chowdhurya, S.-D., 2000, A gradient-based approach
for history-matching of both production and 4D seismic data, paper V-8 presented at the ECMOR
VII, Baveno, Italy, 5-8 September.
Gosselin, O., Aanonsen, S.I., Aavatsmarka, I., Cominelli, A., Gonard, R., Kolasinski, M., Ferdinandi, F.,
Kovacic, L. and Neylon, K., 2003, History matching using time-lapse seismic (HUTS), paper SPE
84464 presented at the Annual Conference and Exhibition, Denver, Colorado, 5-8 October.
Gouveia, W.P., Johnston, D.H., Solberg, A. and Lauritzen, M., 2004, Jotun 4D: characterization of fluid
contact movement from time-lapse seismic and production logging tool data, The Leading Edge
23, 1187-1194.
Grijalba-Cuenca, A., Torres-Verdin, C. and Van der Made, P., 2000, Geostatistical inversion of 3D
seismic data to extrapolate wireline petrophysical variables away from the well, paper SPE 63283
presented at the Annual International Technical Conference, Dallas, Texas, USA, 1-4 October.
Guardiano, F. and Srivastava, R.M., 1993, Multivariate geostatistics: beyond bivariate moments. In:
A. Soares (Editor), Geostatistics-Troia. Kluwer Adademic Publications, Dordrecht, Netherlands,
133-144.
Gunning, J. and Glinsky, M.E., 2007, Detection of reservoir quality using Bayesian seismic inversion,
Geophysics 72, R37-R49.
Haas, A. and Dubrule, O., 1994, Geostatistical inversion a sequential method of stochastic reservoir
modeling constrained by seismic data, First Break 12, 561-569.
239
References
Hadj-Kacem, N. and Pivot, F., 2005, Crafty techniques for populating geomodels with seismic
attributes data, paper P086 presented at the EAGE 67th Conference & Exhibition, Madrid, Spain,
13-16 June.
Hadj-Kacem, N., Berthet, P., Pivot, F. and Samier, P., 2006, Methodology for integration of small scale
3D/4D seismic features in reservoir simulations, paper E032 presented at the EAGE 68th Conference
& Exhibition, Vienna, Austria, 12 -15 June.
Haldorsen, H.H. and Damsleth, E., 1990, Stochastic modeling, Journal of Petroleum Technology,
April, 404-412.
Hicks, G.J. and Francis, A.M., 2007, Prestack stochastic seismic inversion of the Brenda Field, paper
A020 presented at the EAGE 69th Conference & Exhibition, London, UK, 11-14 June.
Hu, L.-Y., 2000, Gradual deformation and iterative calibration of Gaussian-related stochastic models,
Mathematical Geology 32, 87.
Huang, X., Will R., Khan, M. and Stanley, L., 1997, Reservoir characterization by integration of timelapse seismic and production data, paper SPE 38695 presented at the Annual Technical Conference
and Exhibition, San Antonio, TX, USA
Isaaks, E.H. and Srivastava, R.M., 1989, An introduction to applied geostatistics, Oxford University
Press, 561p.
Johann, P., Fournier, F., Souza, O., Eschard, R. and Beucher, H., 1996, 3D stochastic modeling constrained by well and seismic data on a turbidite field, paper SPE 36501 presented at the SPE Annual
Technical Conference and Exhibition, Denver, CO, USA, 6-9 October.
Joseph, C., Fournier, F. and Vernassa, S., 1999, Pseudo-well methodology: a guiding tool for lithoseismic interpretation, 69th Annual International Meeting, SEG Expanded Abstract.
Journel, A.G. and Huijbregts, C. J., 1978, Mining geostatistics, Academic Press, 600p.
Journel, A.G. and Gomez-Hernandez, J.J, 1989, Stochastic imaging of the Wilmington clastic sequence,
SPE 19857.
Jugla, F., Rapin, M., Legeron, S., Magneron, C. and Livingstone, L., 2004, Improving 4D seismic
repeatability using 3D factorial kriging, paper C028 presented at the EAGE 66th Conference &
Exhibition, Paris, France, 7-10 June.
Julien, P., Pivot, F., Douillard, A., El-Ouair, Y. and Toinet, S., 2002, The importance of pre-stack massive seismic modeling for AVO calibration and seismic reservoir characterization, paper presented
at the SEG International Exposition and 72nd Annual Meeting, Salt Lake City, Utah, USA.
Kalkomey, C. T., 1997, Potential risks when using seismic attributes as predictors of reservoir properties, The Leading Edge 16, 247-251.
240
References
Kelkar, M. and Perez, G., 2002, Applied geostatistics for reservoir characterization, SPE, 264p.
Kitanidis, P.K., 1997, Introduction to geostatistics, applications in hydrogeology, Cambridge University
Press, 249p.
Kleemeyer, M., Dankbaar, J.W.M., Staples, R. and Stammeijer, J.G.F., 2006, 4D probabilistic inversion
to detect remaining oil in Troll West oil rim, paper E020 presented at the EAGE 68th Conference &
Exhibition, Vienna, Austria, 12-15 June.
Kretz, V., Le Ravalec-Dupin, M. and Roggero, F., 2004, An integrated reservoir characterization study
matching production data and 4D seismic, SPE Reservoir Evaluation & Engineering 7, 116-122.
Lamy, P., Swaby, P.A, Rowbotham, P.S. and Dubrule, O., 1999, From seismic to reservoir properties with
geostatistical inversion, SPE Reservoir Evaluation and Engineering 2, 334-340.
Landa, J.L. and Horne, R.N., 1997, A procedure to integrate well test data, reservoir performance
history and 4-D seismic information into a reservoir description, paper SPE 38653 presented at the
SPE Annual Technical Conference and Exhibition.
Lecerf, D. and Weisser, T., 2003, New approach for 4D processing with OBC data and marine streamer data, paper A-29 presented at the EAGE 65th Conference & Exhibition, Stavanger, Norway,
2-5 June.
Lee, S.H., Malallah, A. and Datta-Gupta, A., 2000, Multiscale data integration using markov random
fields, paper SPE 63066 presented at the SPE Annual Technical Conference and Exhibition, Dallas,
Texas, 1-4 October.
Leguijt, J., 2001, A promising approach to subsurface information integration, paper L-35 presented
at the EAGE 63rd Conference & Technical Exhibition, Amsterdam, The Netherlands, 11-15 June.
Le Loch, G. and Galli, A., 1996, Truncated plurigaussian method: theoretical and practical point of
view, Geostatistics Wollongong 1, 211-222.
Le Ravalec-Dupin, M., Noetinger, B. and Hu, L.-Y, 2000, The FFT moving average (FFT-MA) generator: an efficient numerical method for generating and conditioning Gaussian simulations,
Mathematical Geology 32, 701.
Le Ravalec-Dupin, M., 2005, Inverse stochastic modelling of flow in porous media, application to reservoir characterization, Editions Technip, 194p.
Lia, O., Omre, H., Tjelmeland, H., Holden, L. and Egeland, T., 1997, Uncertainties in reservoir production forecasts, AAPG Bulletin 81, 775-802.
Lumley, D., Nur, A., Strandenes, S., Dvorkin, J. and Packwood, J., 1994, Seismic monitoring of oil
production: a feasibility study, SEG Expanded Abstract 13, 319.
Lumley, D.E., Behrens, R.A. and Wang, Z., 1997, Assessing the technical risk of a 4-D seismic project,
The Leading Edge 16, 1287-1292.
241
References
MacBeth, C., 2004, A classification for the pressure-sensitivity properties of a sandstone rock frame,
Geophysics 69, 497-510.
MacBeth, C., 2007, An introduction to quantitative 4D seismic interpretation for dynamic reservoir
description, course notes for EAGE Education Days.
Magneron, C., Sandjivy, L. and Vasseur, M., 2006, 3D seismic amplitude filtering using dip steered
geostatistics, paper G037 presented at the EAGE 68th Conference & Exhibition, Vienna, Austria,
12-15 June.
Malallah, A., Perez, H., Datta-Gupta, A. and Alamoudi, W., 2003, Multiscale data integration using
Markov random fields and Markov chain Monte Carlo: a field application in the Middle-East,
paper SPE 81544 presented at the 13th Middle East Oil Show & Conference, Bahrain 9-12 June.
Mallet, J-L, Geomodeling, 2002, Oxford University Press, 599p.
Mallet, J.-L., Moyen, R., Frank, T., Castanie, L., Leflon, B. and Royer, J.-J., 2004, Getting rid of stratigraphic grids, paper B019 presented at the 66th EAGE meeting, Paris, France, 7-10 June.
Marion, D., Insalaco, E., Rowbotham, P., Lamy, P. and Michel, B., 2000, Constraining 3D static
models to seismic and sedimentological data: a further step towards reduction of uncertainties,
SPE65132.
Matheron, G., 1971, The theory of regionalized variables and its applications, Les Cahiers du Centre
de Morphologie Mathmatique de Fontainebleau, no 5, published by the Ecole Nationale Suprieure
des Mines de Paris.
Matheron, G., Beucher, H., de Fouquet, C. and Galli, A., 1987, Conditional simulation of the geometry
of fluvio-deltaic reservoirs, SPE16753.
Mavko, G., Mukerji, T. and Dvorkin, J., 1998, The rock physics handbook, Cambridge University
Press, 329p.
Mavko, G. and Mukerji, T., 1998, A rock physics strategy for quantifying uncertainty in common
hydrocarbon indicators, Geophysics 63, 19972008.
Mazzotti, A. and Zamboni, E., 2003, Petrophysical inversion of AVA data, Geophysical Prospecting 51,
517-530.
Menezes, C. and Gosselin, O., 2006, From logs scale to reservoir scale: Upscaling of the petroelastic
model, paper SPE 10023 presented at the SPE Europec/EAGE Annual Conference and Exhibition,
Vienna, Austria, 12-15 June.
Merletti, G.D. and Torres-Verdn, C., 2006, Accurate detection and spatial delineation of thin-sand
sedimentary sequences via joint stochastic inversion of well logs and 3D prestack seismic amplitude
data, paper SPE 10244 presented at the Annual Technical Conference and Exhibition, San Antonio,
Texas, USA, 24-27 September.
242
References
Mishra, S., 1998, Alternatives to Monte-Carlo simulation for probabilistic reserves estimation and
production forecasting, paper SPE 49313 presented at the Annual Technical Conference and
Exhibition, Denver, USA, 6-9 October.
Morgan, M.G. and Henrion, M., 1990, Uncertainty: a guide to dealing with uncertainty in quantitative
risk and policy analysis, Cambridge University Press, 332p.
Mukerji, T., Jrstad, A., Avseth, P., Mavko, G. and Granli, J.R., 2001a, Mapping lithofacies and
pore-fluid probabilities in a North Sea reservoir: seismic inversions and statistical rock physics,
Geophysics 66, 988-1001.
Mukerji, T., Avesth, P., Mavko, G., Takahashi, I. and Gonzalez, E.F., 2001b, Statistical rock physics:
combining rock physics, information theory, and geostatistics to reduce uncertainty in seismic
reservoir characterization, The Leading Edge 20, 313-319.
Mukerji, T. and Mavko, G., 2005, The flaw of averages and the pitfalls of ignoring variability in rock
physics interpretation, paper presented at the SEG/Houston 75th Annual Meeting.
Mundim, E.C., Johann, P. and Remacre, A.Z., 1999, Factorial kriging analysis: geostatistical filtering
applied to reservoir characterization, The Leading Edge 18,787-788.
Neff, D.B., 1990, Incremental pay thickness modeling of hydrocarbon reservoirs, Geophysics 55,
556-566.
Nivlet, P., Lefeuvre, F. and Piazza, J.L., 2007, 3D seismic constraint definition in deep-offshore turbidite reservoir, Oil & Gas Science and Technology Rev. IFP 62, 249-264.
Odegaard, E. and Avseth, P., 2004, Well log and seismic data analysis using rock physics templates,
First Break 22, October, 37-43.
Oliver, D.S., 1995, Moving averages for Gaussian simulation in two and three dimensions, Mathematical
Geology 27, 939-960.
Omre, H., 1987, Bayesian kriging: merging observations and qualified guesses in kriging, Mathematical
Geology 19, 25-39.
Omre, H. and Halvorsen, B., 1989, The Bayesian bridge between simple and universal kriging,
Mathematical Geology 21, 767-786.
Omre, H., 1992, Heterogeneity models, chapter 6 in J. Kleppe and S. Skjaeveland, eds., SPOR monograph - recent advances in improved oil recovery methods for North Sea sandstone reservoirs,
NPD, Stavanger, Norway, 141-156.
Pyrcz, M.J. and Deutsch, C., 2001, Two artifacts of probability field simulation, Mathematical Geology
3, 775-799.
243
References
Ren, W., Mclennan, J.A., Cunha, L.B. and Deutsch, C.V., 2005, An exact downscaling methodology in
presence of heterogeneity: application to the Athabasca oil sands. Paper SPE PS2005-394 presented
at the International Thermal Operations and Heavy Oil Symposium, Calgary, Alberta, Canada,
1-3 November.
Ringrose, P.S., 2007, Myths and realities in upscaling reservoir data and models, paper D017 presented at the EAGE 69th Conference & Exhibition, London, UK, 11-14 June.
Rowbotham, P., Marion, D., Lamy, P., Insalaco, E., Swaby, P.A.. and Boisseau, Y., 2003, Multidisciplinary
stochastic impedance inversion: integrating geological understanding and capturing reservoir
uncertainty, Petroleum Geoscience, 9, 287-294.
Rubin, Y., 2003, Applied stochastic hydrogeology, Oxford University Press.
Saltzer, R., Finn, C. and Burtz, O., 2005, Predicting Vshale and porosity using cascaded seismic and rock
physics inversion. The Leading Edge 24, 732-736.
Sams, M.S., Atkins, D., Said, P.T., Parwito, E. and van Riel, P., 1999, Stochastic inversion for high resolution reservoir characterization in the Central Sumatra Basin, SPE 57260.
Samson, P., Dubrule, O. and Euler, N., 1996, Quantifying the impact of structural uncertainties on
gross-rock volume estimates, SPE 35535.
Sandjivy, L., 1987, The factorial kriging analysis of regionalised data - Analyse krigeante des donnes
de prospection gochimique PhD dissertation Centre de Gostatistique de Fontainebleau.
Sandjivy, L., Leron, A. and Boudon, S., 2004, The reservoir qualification of seismic data using geostatistics, paper P273 presented at the EAGE 66th Conference & Exhibition, Paris, France, 7-10 June.
Savage, S.L., 2002, The Flaw of Averages, Harvard Business Review, November, 20-21.
Sayers, C.M., Smit, T.J.H., van Eden, C., Wervelman, R., Bachmann, B., Fitts, T., Bingham, J.,
McLachlan, K., Hooyman, P., Noeth, S. and Mandhiri, D., 2003, Use of reflection tomography
to predict pore pressure in overpressured reservoir sands, SEG Technical Program Expanded
Abstracts, 1362-1365.
Sayers, C.M., den Boer, L.D., Nagy, Z.R. and Hooyman, P.J., 2006, Well-constrained seismic estimation
of pore pressure with uncertainty, The Leading Edge 25, 1524-1526.
Scott, D.W., 1992, Multivariate density estimation, John Wiley & Sons, 317p.
Sengupta, M., Mavko, G. and Mukerji, T., 2003, Quantifying subresolution saturation scales from timelapse seismic data: a reservoir monitoring case study, Geophysics 68, 803-814.
Shtuka, A. and Mallet, J.L., 2001, Stochastic simulation of impedances fitting seismic amplitudes and
well data, 21st Annual Gocad Meeting.
Silverman, B.W., 1986, Density estimation for statistics and data analysis, Chapman & Hall, 175p.
244
References
Sinvhal, A. and Khattri, K., 1983, Application of seismic reflection data to discriminate subsurface
lithostratigraphy, Geophysics 48, 1498-1513.
Skjervheim, J.-A., Evensen, G., Aanonsen, S.I., Ruud, B.O. and Johansen, T.A., 2005, Incorporating
4D seismic data in reservoir simulation models using ensemble Kalman filter, paper 95789 SPE
presented at the Annual Technical Conference and Exhibition, Dallas, Texas, USA,
Spikes, K., Mukerji, T., and Dvorkin, J., 2007, Stochastic rock physics inversion for thickness, lithology,
porosity, and saturation, paper to be presented at the SEG International Exposition and 77th Annual
Meeting, San Antonio, Texas.
Srivastava, R.M., 1992, Reservoir characterization with probability field simulation, SPE 24753.
Stephen, K.D., Soldo, J., MacBeth, C. and Christie, M., 2005, Multiple model seismic and production history matching: a case study, paper SPE 94173, Proceedings of the 14th Europec Biennial
Conference Madrid, Spain, 13-16 June.
Stephen, K.D., 2006, Measuring the value of time-lapse (4D) seismic as part of history matching in
the Schiehallion UKCS Field, paper A007 presented at the 10th European Conference on the
Mathematics of Oil Recovery, Amsterdam, The Netherlands.
Strebelle, S.B. and Journel, A.G., 2001, Reservoir modeling using multiple-point statistics, paper SPE
71324 presented at the SPE Annual Technical Conference and Exhibition, New Orleans, Louisiana,
USA.
Strebelle, S., Payrazyan, K. and Caers, J., 2003, Modeling of a deepwater turbidite reservoir conditional to seismic data using principal component analysis and multiple-point geostatistics, SPE
Journal, September, 227-235.
Tarantola, A., 1987, Inverse problem theory: methods for data fitting and model parameter estimation, Elsevier Science Publ., Amsterdam.
Thomas, P., Le Ravalec-Dupin, M. and Roggero, F., 2005, History matching reservoirs models simulated from the Pluri-Gaussian Method, paper SPE 94168 presented at the SPE Europec/EAGE
Annual Conference and Exhibition, Madrid, Spain, 13-16 June.
Toinet, S., 2004, 4D feasibility and calibration using 3D seismic modeling of reservoir models, paper
SPE 88783 presented at the 11th Abu Dhabi International Petroleum Exhibition and Conference,
Abu Dhabi, UAE, 10-13 October.
Tou, J.T. and Gonzalez, R.C., 1974, Pattern recognition principles, Addison-Wesley, 377p.
Tran, T.T., Wen, X.-H. and Behrens, R.A., 1999, Efficient conditioning of 3D fine-scale reservoir
model to multiphase production data using streamline-based coarse-scale inversion and geostatistical downscaling, paper SPE 56518 presented at the Annual Technical Conference and Exhibition,
Houston, Texas, USA.
245
References
Tran, T.T., Deutsch, C.V. and Xie, Y., 2001, Direct geostatistical simulation with multiscale well,
seismic, and production data, paper SPE 71323.
Vasco, D.W., Datta-Gupta, A., Behrens, R., Condon, P. and Rickett, J., 2004, Seismic imaging of
reservoir flow properties: time-lapse amplitude changes, Geophysics 69, 1425-1442.
Veire, H.H., Borgos, H.G. and Landro, M., 2006, Stochastic inversion of pressure and saturation
changes from time-lapse AVO data, Geophysics 71, C81-C02.
Verly, G., 1993, Sequential Gaussian co-simulation: a simulation method integrating several types of
information, Geostatistics Troia 92, Soares ed., Kluwer publishers, Dordrecht, The Netherlands,
543-554.
Waggoner, J., 2001, Integrating time-lapse 3D (4D) seismic data with reservoir simulators, paper F-02
presented at the EAGE 63rd Conference & Technical Exhibition, Amsterdam, The Netherlands,1115 June.
Watts, G.F.T., Jizba, D., Gawith, D.E. and Gutteridge, P.A., 1996, Reservoir monitoring of the Magnus
Field through 4D time-lapse seismic analysis, Petroleum Geoscience, 2, 361-372.
Wen, X.-H., Tran, T., Behrens, R. and Gomez-Hernandez, J., 2002, Production data integration
in sand/shale reservoirs using sequential self-calibration and geomorphing: a comparison, SPE
Reservoir Evaluation & Engineering, 5, 255-265.
Xu, S. and White, R.E., 1995, A new velocity model for clay-sand mixtures, Geophysical Prospecting
43, 91-118.
Xu, W., Tran, T.T., Srivastava, R.M. and Journel, A.G., 1992, Integrating seismic data in reservoir
modeling: the collocated cokriging alternative, SPE 24742.
Xu, W. and Journel, A.G., 1993, GTSIM: Gaussian truncated simulations of reservoir units in a West
Texas carbonate field, SPE 27412.
Xu, W., 1995, Stochastic modeling of reservoir lithofacies and petrophysical properties, PhD
Dissertation, Stanford University.
Yaglom, A.M., 1962, An introduction to the theory of stationary random functions, Prentice-Hall.
Yuh, S.H., Yoon, S., Gibson Jr., R.L. and Datta-Gupta, A., 2000, 4D seismic feasibility study based on an
integrated reservoir model, paper OTC 12135 presented at the Offshore Technology Conference,
Houston, Texas, USA.
Zhu, H. and Journel, A., 1993, Formatting and integrating soft data: stochastic imaging via the
Markov-Bayes algorithm, Geostatistics Troia 92, Soares ed., Kluwer publishers, Dordrecht,
The Netherlands, 1-12.
246
Acknowledgements
Acknowledgements
First, I would like to thank CGGVeritas for allowing me to dedicate time for the preparation of the
course material. Thank you also to the EAGE staff, in particular Anne-Claire Hoenson, Marieke van
Ruijven, Marcel van Loon, Alejandra Reynaldos Rojas and Linda Molenaar for their help in editing the
course book and organizing the lectures tour. I am especially grateful to the following colleagues who
kindly provided illustrations for the course material: Tom Tran and Sebastien Strebelle from Chevron,
Jef Caers and Kyle Spikes from Stanford University, Arild Buland from Statoil, Carlos Torres-Verdin
from the University of Texas at Austin, Frdrik Pivot, Leon Barens and Franois Jugla from Total,
Ran Bachrach from Schlumberger, Hlne Beucher from ENSMP, Guillaume Caumon from ENSG
Nancy, Mickaele Le Ravalec-Dupin and Frdric Roggero from IFP, Ashley Francis from Earthworks
Environment & Resources and Thierry Colou, Rmi Moyen, Jean-Philippe Coulon, Jacques Frelet,
Delphine Hartmann, Raphael Bornard, Thierry Crozat and Fabien Allo from CGGVeritas. I am also
indebted to Lennert den Boer and David Psaila, former colleagues from WesternGeco. Finally, special
thanks to Olivier Dubrule from Total, who carefully reviewed the course notes and provided useful
advice for their preparation.
247
List
of MainMain Mathematical
Mathematical Symbols
List of
Symbols
ci
C(h)
Cxz(h)
Cij
E[]
f(x)
f
f(x|z)
f(x,z)
F(x)
F-1(x)
g
G(x)
h
h
hij
Ip
Is
(i,j,k)
K
mx
m(u)
mx|z
Pe
Po
Pp
p(ci|z)
R(u)
Sw
T
(u,v,w)
ui
uo
v
Vc
Vp
Vs
wi
x(u)
x
xobk
xock
Litho-class i
Covariance between RVs x(u) and x(u+h)
Cross-covariance between RVs x(u) and z(u+h)
Covariance between RVs defined at locations ui and uj
Expected value
PDF of RV x
Generic representation of PEM (Ch 6 and 7 only)
Conditional PDF of RV x given z
Joint PDF of RVs x and z
CDF of RV x
Inverse CDF of RV x
Gravity constant (Ch 6 only)
CDF of Gaussian RV x
Modulus of distance vector or kernel smoothing parameter (Ch 4)
Distance vector
Distance vector between points ui and uj
P-wave acoustic impedance
S-wave acoustic impedance
Indices of one cell in a stratigraphic grid
Kernel function
Mean of RV x
Mean of a RV at location u
Conditional mean of RV x given value of RV z
Effective pressure
Overburden pressure
Pore pressure
Posterior probability for litho-class ci given seismic attribute z
residual at location u
Water saturation
Two-way travel time
Cartesian coordinates of one point in space
Location vector for estimated point
Location vector for estimated point
Cokriging weight for secondary data
Clay volume fraction
P-wave velocity
S-wave velocity
Kriging weight for data point i
RV defined at location u
Estimate of RV x
Block kriging estimate at location uo
Cokriging estimate at location uo
249
xosk
xi
x
V
(h)
x2
2
x|z
ck2
sk2
250
List
of MainMain Abbreviations
Abbreviations
List of
AI
BK
CDF
CCK
CK
DPI
EDK
FFT
FK
FOM
GDM
GI
IK
KDE
LMS
LVM
MA
MAP
MCS
MCMC
ML
MPS
MRF
MSE
OK
PEM
PDF
PFS
PGS
PR
RV
SGS
SGSBK
SIS
SK
TGS
UK
Acoustic Impedance
Bayesian Kriging
Cumulative Distribution Function
Collocated Cokriging
Cokriging
Direct Petrophysical Inversion
External Drift Kriging
Fast Fourier Transform
Factorial Kriging
First Order Moment
Gradual Deformation Method
Geostatistical Inversion
Indicator Kriging
Kernel Density Estimation
Linear Mean Square
Locally Variable Mean
Moving Average
Maximum A Posteriori
Monte Carlo Simulation
Markov Chain Monte Carlo
Maximum Likelihood
Multi-Point Statistics
Markov Random Field
Mean Square Error
Ordinary Kriging
Petro-Elastic Model
Probability Distribution Function
Probability Field Simulation
Pluri-Gaussian Simulation
Poisson Ratio
Random Variable
Sequential Gaussian Simulation
Sequential Gaussian Simulation with Block Kriging
Sequential Indicator Simulation
Simple Kriging
Truncated Gaussian Simulation
Universal Kriging
251
Index
Index
Auto-covariance 42, 59
Anisotropy 33, 35, 36, 37, 39, 41, 46, 48, 49,
67, 69, 70, 123, 125, 131, 135, 231
Bayesian classification 5, 105, 107, 109, 113,
115, 123, 165, 167
Bayesian updating 61, 62, 63, 86, 94, 95, 97,
98, 99, 122, 123, 125, 148, 177
Coefficient of correlation 15, 24, 25, 27, 29,
30, 42, 59, 61, 62, 64, 65, 85, 153
Cokriging 3, 5, 11, 21, 35, 42, 43, 58, 59, 60,
61, 62, 63, 64, 65, 74, 75, 76, 77, 79, 83, 84, 85,
86, 89, 99, 100, 123, 163, 177, 197, 229, 230
Conditional simulation 5, 21, 231
Conditional mean 27, 185
Conditional probability distribution 106
Conditional variance 27, 30
Confusion matrix 118, 119
Correlation length or correlation range 29, 30,
33, 35, 37, 39, 41, 42, 48, 51, 54, 102, 125, 148
Covariance 24, 25, 27, 28, 29, 30, 31, 32, 33,
34, 35, 36, 37, 41, 42, 45, 46, 47, 48, 51, 52,
59, 61, 67, 69, 70, 74, 77, 81, 86, 92, 93, 95,
99, 100, 102, 104, 107, 120, 123, 128, 131, 135,
142, 146, 148, 150, 153, 154, 181, 189
Cross-correlation 40, 41, 42, 195
Cross-covariance 41, 42, 59, 61, 74, 77, 123
253
Index
Sill 33, 39
254
Index
255