Professional Documents
Culture Documents
x ME FEATURE
PODmodels
From Materials Evaluation, Vol. 73, No. 1, pp: 5561.
Copyright 2015 The American Society for Nondestructive Testing, Inc.
Developments in
Probability of
Detection Modeling
and Simulation
Studies
by Jeremy S. Knopp, Frank Ciarallo, and
Ramana V. Grandhi
55
ME FEATURE w
x developments in probability of detection modeling
l
l
l
l
l
1.0
0.8
a90
a90/95
POD
0.6
a50
0.4
0.2
POD (a)
Upper confidence bound
0.0
0
10
12
14
16
a (mm)
Figure 1. Conventional two-parameter probability of detection (POD) model.
56
pi = + ( )
exp ( b0 + b1 log ai )
1 + exp ( b0 + b1 log ai )
pi = + ( ) ( b0 + b1 log ai )
(logit)
(probit)
POD
(a)
large discontinuity sizes. Figure 2c shows a fourparameter model that requires both and to be
calculated.
Depending on whether and are included in the
model, there are four candidate models:
l Two-parameter model that does not include or .
l Three-parameter model with a lower asymptote, .
l Three-parameter model with an upper asymptote, .
l Four-parameter model that includes both and .
Historically, the logit and probit links have been
found to fit POD data well, so there are a total of eight
possible models. The question of which model is
appropriate for a given data set can be answered by
using the bayes factor approach described in prior
work (Kass and Raftery, 1995; Knopp and Zeng, 2013).
Recently, an alternative technique of computing confidence intervals for POD models that include lower and
upper asymptotes via MCMC was introduced (Knopp
and Zeng, 2013). MCMC techniques are very similar to
monte carlo techniques with one important distinction: the samples in monte carlo are statistically independent, which means an individual sample does not
depend on a previous sample. In MCMC, the samples
are correlated with each other. MCMC has proven to
be an effective way to compute multidimensional
integrals that occur in bayesian calculations. Since
MCMC is the computational engine that enables
bayesian analysis, computing confidence bounds with
non-informative priors is a graceful first step to introducing bayesian techniques, which is necessary for
model-assisted POD.
POD
(b)
POD
(c)
Bootstrapping
Current guidance for POD studies assumes that a
linear relationship exist between the explanatory
variable, such as discontinuity size, which is
commonly designated with an a in POD literature,
and the measurement response . Typically, a logarithmic transform will remedy cases where the linear
relationship is not established, but this is not always
the case. Models with additional complexity beyond a
linear model are sometimes necessary for proper
analysis of data. This was the case in the analysis of
data from an inspection of subsurface cracks around
fastener sites using eddy current (Knopp et al., 2012).
The difficulty in these cases is that the procedures for
confidence bound calculations are not developed;
however, a flexible approach called bootstrapping was
demonstrated. Bootstrapping is essentially sampling
with replacement, and it is a very easy technique to
implement for more complicated models and is very
useful for model-assisted POD.
57
ME FEATURE w
x developments in probability of detection modeling
Nonparametric Models
The POD model described in MIL-HDBK-1823A
assumes an S-shaped curve described by two parameters (DOD, 2010). The three-parameter and fourparameter models discussed earlier are modified
versions of that model. An entirely different idea is to
not assume any particular model form, which is
referred to in the literature as a nonparametric model.
A nonparametric model was proposed for POD with
the only assumption being that the POD function is
monotonically increasing with respect to discontinuity
size (Spencer, 2011). This model is useful for many
reasons. First it can be used as a screening model by
comparing the form of the nonparametric model with
the selected parametric model. For example, if the
nonparametric model closely follows a threeparameter model with an upper asymptote, chances
are that the three-parameter model is the best fit. It is
generally useful to see what type of model form the
data dictates before forcing a parametric model on it.
Box-cox Transformation
It is always advantageous to use the measured
response data for POD evaluation since there is more
information contained in that form rather than
hit/miss; however, real inspection data often violate
core assumptions required to use a POD model fit.
Another development is the use of the box-cox transformation to mitigate violations of homoscedasticity
for data analysis. Homoscedasticity means that the
scatter in the observations is constant for the discontinuity size range. For cases where there is a relationship between the mean response and the variance,
the box-cox transformation is used to stabilize the
variance. This technique assumes that the relationship
between the error variance, s2i , and mean response,
i, can be described with a power transformation on
in the form of Equation 3. The new regression model
in Equation 4 includes the additional parameter,
which also needs to be estimated.
(3)
(4)
a = a
a i = 0 + 1 ai + i
(5)
gi =
(6)
gi = c ln ai 1 , = 0
where
n is the total number of observations,
c = (i)1/n, which happens to be the geometric
mean of the observations.
Once these standardized observations are
obtained, they are then regressed on a, which in this
case is crack length, and then the sum of squares
error (SSE) is obtained. The optimization problem is
formulated such that the objective is to minimize SSE
with l as a single parameter to be adjusted. An
example of how the box-cox transform is used was
presented in the context of an eddy current inspection
of cracks around fastener sites in an aircraft structure
(Knopp et al., 2012).
Sample Size
One of the more common questions asked about POD
studies is how many samples will be required. Both
the sample size and distribution of the samples will
affect the POD evaluation. In MIL-HDBK-1823A, it is
recommended that there be at least 40 samples when
the signal response data is used and 60 samples for
hit/miss (DOD, 2010). The question of the range of
discontinuity sizes and the distribution of discontinuity sizes is not discussed in MIL-HDBK-1823A, and
has not been examined extensively in the literature
except in a few cases (Berens and Hovey, 1985;
Safizadeh et al., 2004). Recently, this question for
hit/miss data was investigated via simulation and
looked at discontinuity size distribution and the
effects of moving the center of the sample distribution
relative to the true a50 value (Annis, 2014). The recommendations from this study include using a uniform
distribution of discontinuity sizes and that the range
should be from 3 to 97% POD. The number of
specimens agrees with MIL-HDBK-1823A in that a
minimum of 60 specimens should be used.
Bayesian Design of Experiments
A bayesian approach to planning hit/miss experiments
has also been presented (Koh and Meeker, 2014).
This allows engineers to use any prior information that
may be known about a POD curve to assist in
designing the experiment. The conclusion of this work
was that optimal test plans developed purely from
bayesian techniques may not be practical, but they
Simulation Studies
Going forward, the authors recommend simulation
studies to provide the NDT practitioner with a connection between the intuition gained with inspection
experience and the statistical techniques used to
quantify the capability of an inspection. For example,
simulation studies can be used to benchmark the
effect of sample size on confidence bound calculations by means of evaluating probability of coverage.
Probability of coverage is defined as the probability
that the interval contains the quantity of interest. In
this work, covering a90 with a 95% upper confidence
bound (that is, a90/95) is of particular interest, so the
probability that a90/95 is greater than the true value of
a90 as defined in Equation 7 is what is meant by probability of coverage in this paper. The objective of simulation studies is to show how often (in terms of
percent) a confidence interval contains the true
parameter of interest.
(7)
0.6
0.5
0.4
0.2
0.1
0.0
0
(8)
a = 0 + 1 a (1 + mult ) + add
Figure 3. Simulated data with linear model form that includes additive and
multiplicative noise.
1.0
0.9
0.8
True POD
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.0
0
a (mm)
0.3
a (mm)
Figure 4. True probability of detection (POD) curve available from
100 000 simulated observations.
59
ME FEATURE w
x developments in probability of detection modeling
12
10
Range
8
6
4
True a90
2
0
a90
(30 obs)
a90/95
(30 obs)
a90
(100 obs)
a90/95
(100 obs)
Probability of detection
Figure 5. Box plots for a90 and a90/95 values for 30 and 100 observations (obs) for
two-parameter hit/miss model.
60
61