Professional Documents
Culture Documents
s
2
product
s
2
assay
Bias
2
_
Process capability, in turn, is related to the percentage of
measurements that are likely to fall outside of the spe-
cifications. Thus acceptance criteria on the amount of
bias and analytical variability can be established based
Assay Validation 65
A
D
o
w
n
l
o
a
d
e
d
B
y
:
[
U
n
i
v
e
r
s
i
t
y
o
f
A
l
b
e
r
t
a
]
A
t
:
0
6
:
3
5
7
J
a
n
u
a
r
y
2
0
0
9
upon the knowledge of the product variability and the
desired process capability.
When the limits have not been established for a
particular product characteristic, such as during the early
development of a drug or biologic, acceptance criteria for
an assay attribute might be specified on the basis of a
typical expectation for the analytical method as well as on
the nature of the measurement in the particular sample.
Thus for example, high-performance liquid chromato-
graphy for the active compound in the final formulation of
a drug might be typically capable of yielding measure-
ments equal to or less than 10% RSD, while an immu-
noassay used to measure antigen content in a vaccine
might only be capable of achieving up to 20% RSD.
Measurement of a residual or an impurity by either me-
thod, on the other hand, may only achieve up to 50%
RSD, owing to the variable nature of measurement at low
analyte concentrations.
VALIDATION DATA ANALYSIS
Analysis of Parameters Related to Accuracy
Analytical accuracy and linearity can be parameterized as
a simple linear model, assuming either absolute or relative
bias. In one model, if we let m represent the known analyte
content and x is the measured amount, then the accuracy
can be expressed as x=m at a single concentration, or as
x=a+bm, where a=0 and b=1, across a series of
concentrations. This linear equation can be rewritten as
x m a b 1m
In this parameterization, a is related to the accuracy of the
analytical method and represents the constant bias in an
assay measurement (usually reported in the units of
measurement of the assay), while (b1) is related to the
linearity of the analytical method and represents the
proportional bias in an assay measurement (usually
reported in the units of measurement in the assay per
unit increase in that measurement, e.g., 0.02 mg per
microgram increase in content). Data from a spiking
experiment can be utilized to estimate a and b1, and
these (with statistical confidence limits if the validation
experiment has been designed to establish statistical con-
formance) can be compared with the acceptance cri-
teria established for these parameters.
To illustrate, consider the example in Table 2. A vali-
dation study has been performed in which samples have
been prepared with five levels of an analyte ([C] in mg),
and their content is determined in six runs of an assay.
These data are depicted graphically, along with the results
from their analysis, in Fig. 1. A linear fit to the data yields
the following estimates of intercept (note that the data
have been centered in order to obtain a centered
intercept) and slope:
^a
0
0:04 0:08; 0:01;
^
b 0:81 0:70; 0:85
The slope can be utilized to estimate the bias per unit
increase in drug concentration:
^
b 1 0:19 0:30; 0:15
This proportional bias is significant, owing to the fact that
the confidence interval excludes 0. It is more desirable,
however, to react to the magnitude of the bias and to
judge that there is an unacceptable nonlinearity when the
bias is in excess of a prespecified acceptance limit. Thus
there is a 0.2-mg decrease per unit increase in concen-
tration (as much as 0.3 mg per unit decrease in
concentration based on the confidence interval). If the
range in the concentration of the drug is typically 46 mg
(i.e., a 2-mg range), this would predict that the bias as a
result of nonlinearity is 0.4 mg (0.6 mg in the
confidence interval). This can be judged to be of con-
sequence or not in testing product; for example, if the
specification limits on the analyte are 51 mg/dose,
much of this range will be consumed by the proportional
Table 2 Example of results from a validation experiment
in which five levels of an analyte were tested in six runs of
the assay
[C] 1 2 3 4 5 6 Avg.
3 3.3 3.4 3.2 3.1 3.4 3.3 3.3
4 4.2 4.2 4.2 4.4 4.0 4.3 4.2
5 5.0 4.9 4.9 5.0 4.8 4.9 4.9
6 5.8 5.8 5.6 5.9 5.9 5.7 5.8
7 6.4 6.7 6.7 6.8 6.6 6.4 6.6
Fig. 1 Validation results from an assay demonstrating
proportional bias, with low-concentration samples yielding
high measurements and high-concentration samples yielding
low measurements.
66 Assay Validation
D
o
w
n
l
o
a
d
e
d
B
y
:
[
U
n
i
v
e
r
s
i
t
y
o
f
A
l
b
e
r
t
a
]
A
t
:
0
6
:
3
5
7
J
a
n
u
a
r
y
2
0
0
9
bias, potentially resulting in an undesirable failure rate in
the measurement of that analyte.
When the purified analyte is unavailable for spiking,
and the levels of the analyte in the validation have been
attained through dilution, then the data generated from
that series can be alternatively evaluated using the model
x = am
b
(note that m can represent either the expected
concentration upon dilution or the actual dilution). Taking
the logs, the data from the dilution series is used to fit the
linear model,
logx loga b logm
The intercept of this model has no practical meaning,
while the proportional bias (sometimes called dilution
effect) can be estimated as 2
b 1
1 if m is in units of
concentration and as 2
b + 1
1 if m is in units of dilution
(note that the units of dilution effect are percent bias
per dilution increment; the dilution increment used here
is twofold).
To illustrate, consider the example in Table 3. A
dilution series of a sample is performed in each of the five
runs of an assay. These data are depicted graphically,
along with the results from their analysis, in Fig. 2. A
linear fit of log-titer to log-dilution yields an estimated
slope equal to
^
b 1:01 1:04; 0:98. The corres-
ponding dilution effect is equal to 2
1.01 + 1
1= 0.007
(i.e., a 0.7% decrease per twofold dilution).
The dilution effect can also be calculated as a function
of the estimated slopes for the standard (slope as a
function of dilution rather than concentration) and the
test sample:
Dilution effect 2
1
^
b
T
=
^
b
S
with the corresponding standard error a function of the
standard error of a ratio:
SE
^
b
T
=
^
b
S
1
^
b
2
S
VAR
^
b
T
^
b
T
=
^
b
S
2
VAR
^
b
S
_ _
Va
21
a
2
21
Va
11
a
2
11
2
cova
21
; a
11
a
21
a
11
_ _
^
b
2
C
where
Va
i1
l
1
l
2
n l
2
l
1
2
_ _
a
2
i2
and
cova
21
; a
11
l
1
l
2
n l
2
l
1
2
_ _
a
12
a
22
Table 3 Example of results from a validation experiment in
which five twofold dilutions of an analyte were tested in five
runs of the assay
Dilution 1 2 3 4 5 Titer
1 75 90 93 79 72 81
2 43 45 46 35 40 83
4 20 23 22 18 21 83
8 11 10 12 10 10 85
16 4 5 6 4 5 78
Fig. 2 Results from a validation series demonstrating satisfact-
ory linearity, i.e., approximately a twofold decrease in response
with twofold dilution.
Assay Validation 67
A
D
o
w
n
l
o
a
d
e
d
B
y
:
[
U
n
i
v
e
r
s
i
t
y
o
f
A
l
b
e
r
t
a
]
A
t
:
0
6
:
3
5
7
J
a
n
u
a
r
y
2
0
0
9
The concordance slope can be used, along with the
standard error of that estimate, to measure the discordance
of the experimental analytical method relative to the
referee method and its corresponding confidence
interval:
Discordance 2
b
C
1
This is similar to the dilution effect discussed earlier
and represents the percentage difference in reported
potency per twofold increase in result as measured in
the referee method.
When the two assays are scaled alike, the accuracy
of experimental method relative to the referee me-
thod can be assessed through a simple paired variate
analysis. When the percent difference is expected to be
constant across the two methods, the data are analyzed
on a log scale. An estimate of the simple paired dif-
ference in the logs can be used, in conjunction with the
standard error of the paired difference, to estimate the
percentage difference in measurements obtained from
the experimental procedure relative to the referee
method:
[6]
d t
n1
s
d
=
n
p
Percent difference 100e
d
1%
Note that as usual, the significance of the percentage
difference should be judged primarily on the basis of
practical considerations rather than statistical significance.
The specificity of an analytical procedure can be
assessed through an evaluation of the results obtained
from a placebo sample (i.e., a sample containing all
constituents except the analyte of interest) relative to
background. A statistical evaluation might be composed
of either a comparison of instrument measurements
obtained for this placebo, with background measure-
ments (i.e., through background-corrected placebo
measurements), or from measurements made on the
placebo alone. In either case, an increase over the
detection level of the assay would indicate that some
constituent of the sample, in addition to analyte, is con-
tributing to the measurement in the sample. This should
be assessed by noting that the upper bound on a con-
fidence interval falls below some prespecified level in
the assay.
Analysis of Parameters Related to Variability
Estimates of repeatability (intrarun, or within-run, vari-
ability) and precision (interrun, or between-run, variabil-
ity) can be established by performing a variance com-
ponent analysis on the results obtained from the data
generated to assess accuracy and linearity.
[7]
Other
sources of variability (e.g., laboratories, operators, and
reagent lots) can be incorporated into the pattern of runs
performed during the validation to estimate the effects of
these sources of variability. The expected mean squares
are shown in Table 4. From this, ^ s
2
withinrun
mean square
error (MSE) and ^ s
2
betweenrun
mean square for runs
(MSR)MSE=5 (note that the MSE is a composite of
pure error and the interaction of run and level; pure error
could be separated from the interaction by performing
replicates at each level; however, the within-run variabil-
ity will still be reported as the composite of pure error
and the level-by-run interaction). The total variability
associated with testing a sample in the assay can be
calculated from these variance component estimates as:
^ s
2
Total
^ s
2
B
r
^ s
2
W
nr
Note that this is the variance of y, where r represents the
number of independent runs performed on the sample
and n represents the number of replicates within a run.
A confidence interval can be established on the total
variability as follows:
[8]
^ s
2
Total
^ s
2
B
r
^ s
2
W
nr
1
Jr
_ _
MSR
n J
Jnr
_ _
MSE
c
1
MSR c
2
MSE
q
c
q
S
q
where J represents the number of replicates of a sample
tested in each run of the assay validation experiment (here
J=5 for the number of levels of the sample). The 1 2a
confidence interval on ^ s
2
Total
becomes:
^ s
2
Total
q
G
2
q
c
2
q
S
4
q
; ^ s
2
Total
q
H
2
q
c
2
q
S
4
q
_
_
_
_
Table 4 Analysis of variance table showing expected mean
squares as functions of intrarun and interrun components
of variability
Effect df MS
Estimated
mean square F
Level 4 Mean square
for levels
(MSL)
^ s
2
W
QL
Run 5 MSR ^ s
2
W
5^ s
2
B
MSR/MSE
Error 20 MSE ^ s
2
W
68 Assay Validation
D
o
w
n
l
o
a
d
e
d
B
y
:
[
U
n
i
v
e
r
s
i
t
y
o
f
A
l
b
e
r
t
a
]
A
t
:
0
6
:
3
5
7
J
a
n
u
a
r
y
2
0
0
9
where
G
q
1
1
F
a;df
q
;1
; H
q
1
F
1a;df
q
;1
1;
F
a;df
q
;1
; F
1a;df
q
;1
represent the percentiles of the F-distribution.
The LOD and the LOQ are calculated using the
standard deviation of the responses (^ s) and the slope
of the calibration curve (
^
b) as LOD 3:3^ s=
^
b (note,
the 3.3 is derived from the 95th percentile of the stand-
ard normal distribution, 1.64; here, 2 1.64 = 3.3,
thereby limiting both the false-positive and false-
negative error rates to be equal to or less than 5%) and
LOQ 10^ s=
^
b (note, the 10 corresponds to the restriction
of no more than 10% RSD in the measurement variability;
RSD=100 ^ s=x < 0:10 gives x; thus if the restriction
was changed to not more than 20% RSD, then the
LOQ could be calculated as LOQ 5^ s=
^
b).
These estimates do not account for the calibration
curve, and they can be improved upon by acknowledging
the variability in prediction from the curve.
[9]
For a linear
fit, the LOD is depicted graphically in Fig. 3. The LOQ
can be calculated as prescribed in the ICH guideline,
solving for x
Q
in the following equation:
LOQ 10 ^ s
1
1
n
x
Q
x
2
SXX
_
_
_
_
_
_
The limit of quantitation is depicted graphically in Fig. 4.
The LOQ is more complicated to calculate from a
nonlinear calibration function, such as the four-parameter
logistic regression equation, and involves the first-or-
der Taylor series expansion of the solution equation for
x
Q
[10]
(see the entry on Assay Development). The calcu-
lation of the LOQ for this equation is illustrated in Fig. 5.
FOLLOW-UP
The information from the analytical method validation
can be utilized to establish a testing format for the
samples in the assay. In particular, variance component
estimates of the within-run and between-run variabilities
Fig. 3 Determination of the LOD using the linear fit of the
standard curve plus restrictions on the proportions of false
positives and false negatives.
Fig. 4 Determination of the LOQ using the linear fit of the
standard curve and a restriction on the variability (% RSD) in
the assay.
Fig. 5 Quantifiable range determined as limits within which
assay variability (% RSD) is less than or equal to 20%.
Table 5 Predicted variability (% RSD) for a given number of
runs (r) and a given number of replicates within each run (n)
r/n 1 2 3 1111
1 6% 5% 5% 5.0%
2 4% 4% 4% 3.5%
3 3% 3% 3% 2.9%
Assay Validation 69
A
D
o
w
n
l
o
a
d
e
d
B
y
:
[
U
n
i
v
e
r
s
i
t
y
o
f
A
l
b
e
r
t
a
]
A
t
:
0
6
:
3
5
7
J
a
n
u
a
r
y
2
0
0
9
can be used to predict the total variability associated
with a variety of testing formats. For example, consider
the case where component estimates have been deter-
mined to be equal to ^ s
B
0:05 (5%) and ^ s
W
0:03
(3%). Letting r represent the number of independent runs
that will be performed on a sample, and n be the number
of replicates within a run, the variabilities are listed in
Table 5. Suppose that, because of the process capability,
the analytical variability must be restricted to RSD3%.
As can be observed from the table, this cannot be ac-
complished using a single run of the assay; it must
therefore be achieved through a combination of multiple
runs with multiple replicates within each run.
ANALYTICAL METHOD QUALITY CONTROL
Assay validation does not end with the formal validation
experiment but is continuously assessed through assay
quality control. Control measures of variability and bias
are utilized to bridge to the assay validation and to
warrant the continued reliability of the procedure.
Extra variability in replicates can be detected utilizing
standard statistical process control metrics;
[4]
detection
and prespecified action upon detection can help maintain
the variability characteristics of an assay. For example,
when the measurement of an analyte is obtained from
multiple runs of an assay, those results can be examined
by noting the range (either log(max) log(min) or
max min, depending upon whether the measure-
ments are scaled proportionally or absolutely). A bound
on the range can be derived:
Range ^ s
dimension
d
2
kd
3
where ^ s
dimension
is the variability in the dimension of rep-
lication (here, run to run), d
2
and d
3
are the factors
used to scale the standard deviation to a range, and k is
a factor representing the desired degree of control (e.g.,
k = 2 represents approximately 95% confidence). Hav-
ing detected extra variability, additional measurements
can be obtained, and a result for the sample can be
determined from the remainder of the measurements
made after the maximum and minimum results have been
eliminated from the calculation. When extra variabil-
ity has been identified in instrument replicates of a
standard concentration or a dilution of a test sample, that
concentration of the standard or dilution of the test sample
might be eliminated from subsequent calculations.
Measurements on control sample(s) help identify shifts
in scale and therefore bias in a run of an assay. In-
formation in multiple control samples can typically be
reduced to one or two meaningful scores, such as an
average and/or a slope. These can help isolate problems
and provide information regarding the cause of a
deviation in a run of an assay. Suppose, for example,
that three controls of various levels are used to monitor
shifts in sensitivity in the assay. The average can be used
to detect absolute shifts, because of some underlying
source of constant variability, such as inappropriate
preparation of standards or the introduction of significant
background. The slope can be used to detect proportional
shifts as a result of some underlying source of pro-
portional variability.
Suitability criteria on the performance of column
peaks can detect degeneration of a column, while cha-
racteristics of the standard curve, such as the slope or
EC
50
(EC
50
corresponds to C in a four-parameter logistic
regression) help predict an assay with atypical or
substandard performance.
Care must be taken, however, to avoid either mean-
ingless or redundant controls. The increased false-positive
rate inherent in the multiplicity of a control strategy must
be considered, and a carefully contrived quality control
scheme should acknowledge this.
CONCLUSION
A carefully conceived assay validation should follow
assay development, and precede routine implementation
of the method for product characterization. The purpose
of the validation is to demonstrate that the assay is
fundamentally reliable, and possesses operating charac-
teristics that make it suitable for use. The statistical
estimates collected during the validation can also be used
to design a test plan, that provides the maximum relia-
bility for the minimum effort in the laboratory. Most
importantly, validation of the assay does not end with the
documentation of the validation experiment, but should
continue with assay quality control. The proper choice of
control samples and control parameters helps assure the
continued validity of the procedure during routine use in
the laboratory.
REFERENCES
1. USP XXII. General Chapter 1225, Validation of Compen-
dial Methods; 19821984.
2. Federal Register. International Conference on Harmoniza-
tion; Guideline on Validation of Analytical Procedures:
Definitions and Terminology; March 1, 1995; 11259
11262.
70 Assay Validation
D
o
w
n
l
o
a
d
e
d
B
y
:
[
U
n
i
v
e
r
s
i
t
y
o
f
A
l
b
e
r
t
a
]
A
t
:
0
6
:
3
5
7
J
a
n
u
a
r
y
2
0
0
9
3. International Conference on Harmonization; Guideline on
Validation of Analytical Procedures: Methodology;
November 6, 1996.
4. Montgomery, D. Introduction to Statistical Quality Con-
trol, 2nd Ed.; Wiley: New York, 1991; pp. 203206, 208
210.
5. Morrison, D. Multivariate Statistical Methods; McGraw-
Hill: New York, 1967; 221258.
6. Snedecor, G.; Cochran, W. Statistical Methods, 6th Ed.;
Iowa State Univ. Press: Ames, IA, 1967; 91100.
7. Winer, B. Statistical Principles in Experimental Design,
2nd Ed.; McGraw-Hill: New York, 1971; 244251.
8. Burdick, R.; Graybill, F. Confidence Intervals on Variance
Components; Marcel Dekker: New York, 1992; 2839.
9. Oppenheimer, L.; Capizzi, T.; Weppelman, R.; Mehta, H.
Determining the lowest limit of reliable assay measure-
ment. Anal. Chem. 1983, 55, 638643.
10. Morgan, B. Analysis of Quantal Response Data; Chapman
& Hall: New York, 1992; 370, 371.
11. Massart, D.L.; Dijkstra, A.; Kaufman, L. Evaluation and
Optimization of Laboratory Methods and Analytical
Procedures; Elsevier: New York, 1978.
12. Currie, L.A. Limits for qualitative detection and quanti-
tative determination: Application to radiochemistry. Anal.
Chem. 1968, 40, 586593.
13. Cardone, M.J. Detection and determination of error in
analytical methodology. Part I. The method verification
program. J. Assoc. Off. Anal. Chem. 1983, 66, 1257
1281.
14. Cardone, M.J. Detection and determination of error in
analytical methodology. Part II. Correction for cor-
rigible systematic error in the course of real sample
analysis. J. Assoc. Off. Anal. Chem. 1983, 66, 1283
1294.
15. Miller, J.C.; Miller, J.N. Statistics for Analytical Chem-
istry, 2nd Ed.; Wiley: New York, 1988.
16. Caulcutt, R.; Boddy, R. Statistics for Analytical Chemists;
Chapman & Hall: New York, 1983.
Assay Validation 71
A
D
o
w
n
l
o
a
d
e
d
B
y
:
[
U
n
i
v
e
r
s
i
t
y
o
f
A
l
b
e
r
t
a
]
A
t
:
0
6
:
3
5
7
J
a
n
u
a
r
y
2
0
0
9