You are on page 1of 6

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS—II: ANALOG AND DIGITAL SIGNAL PROCESSING, VOL. 47, NO.

2, FEBRUARY 2000 151

REFERENCES Neural-Network Based Analog-Circuit Fault Diagnosis


[1] E. Hunt, Artificial Intelligence. New York: Academic, 1975. Using Wavelet Transform as Preprocessor
[2] P. Winston, Artificial Intelligence. Reading, MA: Addison-Wesley,
1977. Mehran Aminian and Farzan Aminian
[3] B. G. Batchelor, Pattern Recognition. New York: Plenum, 1978.
[4] R. Gnanadesikan, Methods for Statistical Data Analysis of Multivariate
Observations. New York: Wiley, 1977. Abstract—We have developed an analog-circuit fault diagnostic system
[5] T. M. Cover and P. E. Hart, “Nearest neighbor pattern classification,” based on backpropagation neural networks using wavelet decomposition,
IEEE Trans. Inform. Theory, vol. IT-13, pp. 21–27, Jan 1967. principal component analysis, and data normalization as preprocessors.
[6] R. O. Duda and P. E. Hart, Pattern Classification and Scene Anal- The proposed system has the capability to detect and identify faulty com-
ysis. New York: Wiley, 1973. ponents in an analog electronic circuit by analyzing its impulse response.
[7] J. C. Bezdek, Pattern Recognition with Fuzzy Objective Function Algo- Using wavelet decomposition to preprocess the impulse response drastically
rithms. New York: Plenum, 1981. reduces the number of inputs to the neural network, simplifying its archi-
[8] J. M. Keller, M. R. Gray, and J. A. Givens Jr., “A fuzzy -nearest tecture and minimizing its training and processing time. The second pre-
neighbor algorithm,” IEEE Trans. Syst., Man, Cybern., vol. SMC-15, processing by principal component analysis can further reduce the dimen-
pp. 580–585, July/Aug. 1985. sionality of the input space and/or select input features that minimize diag-
[9] E. Seevinck, R. F. Wassenaar, and H. C. K. Wong, “A wide-band tech- nostic errors. Input normalization removes large dynamic variances over
nique for vector summation and rms-dc conversion,” IEEE J. Solid-State one or more dimensions in input space, which tend to obscure the relevant
Circuits, vol. SC-19, pp. 311–318, June 1984. data fed to the neural network. A comparison of our work with [1], which
[10] O. Landolt, E. Vittoz, and P. Heim, “CMOS selfbiased Euclidean dis- also employs backpropagation neural networks, reveals that our system re-
tance computing circuit with high dynamic range,” Electron. Lett., vol. quires a much smaller network and performs significantly better in fault
28, no. 4, pp. 352–354, Feb. 1992. diagnosis of analog circuits due to our proposed preprocessing techniques.
[11] S. I. Liu and C. C. Chang, “A CMOS square-law vector summation cir-
cuit,” IEEE Trans. Circuits Syst. II, vol. 43, pp. 520–524, July 1996. Index Terms—Analog circuits, electronic circuits, fault diagnosis, neural
[12] C. Y. Huang and B. D. Liu, “Current-mode multiple input maximum networks, wavelet transform.
circuit for fuzzy logic controller,” Electron. Lett., vol. 30, no. 23, pp.
1924–1925, Nov. 1994.
[13] C. Y. Huang, C. Y. Chen, and B. D. Liu, “Current-mode linguistic hedge I. INTRODUCTION
circuit for adaptive fuzzy logic controllers,” Electron. Lett., vol. 31, no.
17, pp. 1517–1518, Aug. 1995. Fault diagnosis in digital electronic circuits has been successfully de-
[14] C. Y. Chen, C. Y. Huang, and B. D. Liu, “Current-mode fuzzy linguistic veloped to the point of automation. However, the process of developing
hedge circuit—Contrast intensification,” in Proc. 1996 IEEE Int. Symp.
test strategies for analog circuits still relies heavily on the engineer’s
Circuits and Systems, Atlanta, GA, May 1996, pp. 511–514.
[15] C. Y. Chen, C. Y. Huang, B. D. Liu, and T. J. Su, “A current-mode fuzzy experience and intuition. This requires the engineer to have some de-
linguistic hedge circuit—More or less,” in Proc. 5th IEEE Int. Conf. tailed knowledge of the circuit’s operational characteristics and expe-
Fuzzy Systems, New Orleans, LA, Sept. 1996, pp. 1080–1085. rience in developing test strategies. As a result, analog fault detection
[16] C. Y. Chen, C. Y. Huang, J. Y. Tsao, and B. D. Liu, “A current-mode and identification is still an iterative and time-consuming process. The
circuit for Euclidean distance calculation,” in Proc. 1997 Int. Symp. VLSI
Technology, Syst.ems, and Applications, Taipei, Taiwan, June 1997, pp.
current state in electronic circuit manufacturing has introduced analog
83–86. and analog/digital hybrid circuits, where the total circuit under test is
[17] C. Y. Chen, C. Y. Huang, and B. D. Liu, “Current-mode defuzzifier cir- large. As a result, there is an urgent need to find a systematic approach
cuit to realize the centroid strategy,” Proc. Inst. Elect. Eng.—Circuits to automate the fault diagnostic process in these circuits where intu-
Devices and Systems, vol. 144, no. 5, pp. 265–271, Oct. 1997.
ition and experience may no longer be sufficient [2], [3].
[18] C. Y. Huang, C. Y. Chen, and B. D. Liu, “Current-mode fuzzy linguistic
hedge circuits,” Analog Integr. Circuits Signal Processing, vol. 19, no. The engineering community began to look into analog test problems
3, pp. 255–278, June 1999. in the mid 1970’s, since analog systems were among the most unreli-
[19] K. Bult and H. Wallinga, “A class of analog CMOS circuits based on able and least testable systems. As a result, many works appeared in
the square-law characteristic of an MOS transistor in saturation,” IEEE the literature and analog fault diagnosis became an active area of re-
J. Solid-State Circuits, vol. SC-22, pp. 357–364, June 1987.
[20] E. Seevinck and R. J. Wiegerink, “Generalized translinear circuit prin- search. During the past few years, there has been significant research on
ciple,” IEEE J. Solid-State Circuits, vol. SSC-26, pp. 1098–1102, Aug. analog fault diagnosis at the system, board, and chip level [1], [3]–[5].
1991. A survey of the research conducted in this area clearly indicates that
[21] R. J. Wiegerink, Analysis and Synthesis of MOS Translinear Cir- analog fault diagnosis is complicated due to the poor fault models, com-
cuits. Boston, MA: Kluwer, 1993.
ponent tolerances, and nonlinearity issues. These difficulties make the
[22] S. Franco, Design with Operational Amplifiers and Analog Integrated
Circuits. New York: McGraw-Hill, 1988, pp. 58–63. application of neural networks to these problems very appealing. In
this approach, all the faults are modeled by a unique set of features
(signatures) that the network learns during the training phase. These
features, together with the associated fault classes, are presented to the
network as input–output pairs. The network is then allowed to adjust
its weight and bias parameters to learn the desired input–output rela-
tionship. Next, the network is presented with a set of features as input
during the testing phase and determines the fault class. This is called
generalization, where the network predicts the fault class based on the
knowledge acquired during the training phase. Fault diagnosis based on

Manuscript received December 1998; revised September 1999. This paper


was recommended by Associate Editor P. Thiran.
M. Aminian is with St. Mary’s University, San Antonio, TX 78228 USA.
F. Aminian is with Trinity University, San Antonio, TX 78212 USA.
Publisher Item Identifier S 1057-7130(00)01464-6.

1057-7130/00$10.00 © 2000 IEEE


152 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS—II: ANALOG AND DIGITAL SIGNAL PROCESSING, VOL. 47, NO. 2, FEBRUARY 2000

neural networks bypasses the difficulties associated with analog fault


diagnosis described earlier, since no physical models of the faults or
tolerance problems are required. Instead, the faulty circuit is simulated
or built and the appropriate features associated with the fault under con-
sideration are extracted from the circuit’s output(s).
There are three recent works that discuss applications of neural net-
works to analog fault diagnosis [1], [2], [6]. Among these, [2] is a
very introductory work that mainly discusses the potential of neural
networks in analog system test development. Reference [1] applies Fig. 1. Hierarchial decomposition of a signal into approximations and details.
neural networks to fault diagnosis of linear circuits without any prepro- Note that each approximation or detail component is further decomposed into its
cessing of the circuit’s impulse response. This can lead to large neural own approximation and detail levels. Here, only two levels of decompositions
network architectures even for relatively small circuits. For a larger are shown. Also shown is the downsampling effect in wavelet analysis.
circuit, the authors attempt two ambiguity groups to keep the clas-
sification problem and neural network size manageable. In this case,
each ambiguity group is assumed to be the collection of seven to eight In this equation, a and b define the degree of scaling and shifting of
faulty components. Even though faults can be indistinguishable if they the mother wavelet (x) ; respectively. The coefficients of expansion
create similar outputs, it is possible to further break down the ambi- C (a; b) for a particular signal can be expressed as
guity groups in this reference through appropriate preprocessing of the
circuit’s output(s) as shown later. The work in [6] also does not involve
preprocessing to reduce the number of input features to a manageable C (a; b) = h a;b x
( )
size. Feature selection is performed in frequency domain and involves
I (x)i = p 1
I (x) x 0 b dx: (2)
a significant number of features, even for a small circuit. Moreover, the
selection process in the authors’ approach requires significant experi-
a a
ence and intuition about circuit’s operational characteristics. These coefficients give a measure of how closely correlated the mod-
Neural-network based fault diagnosis is applicable to analog circuits ified mother wavelet [in (1)] is with the input signal. In (2), I (x) rep-
as long as the faults under consideration result in classes which can be resents the signal and the integration is performed over all possible x
separated from each other and the no-fault class by means of appro- values. Wavelet analysis in its discrete (dyatic) form assumes a = 2j
priate features. Within the context of this work, appropriate features and b = k2j = ka; where (j; k) 2 Z 2 : Approximations and details
refer to the impulse response of the electronic circuit preprocessed by mentioned earlier correspond to high-scale (low-frequency) and low-
wavelet decomposition, principal component analysis (PCA), and nor- scale (high-frequency) components of a signal, respectively. Therefore,
malization. We will show later that through these preprocessing steps, wavelet decomposition can be considered a process that generates a
we can significantly reduce the number of inputs and select appro- family of hierarchically organized low- and high-frequency compo-
priate features such that the neural network architecture required for nents of a signal. At each level, say j; one can construct the approx-
fault diagnosis has a minimal size. In this work, we will concentrate imation Aj and detail Dj of the original signal. The block diagram of
on faults resulting from components having values out of the range al- this decomposition is shown in Fig. 1. One can, therefore, consider ap-
lowed by component tolerances so that we can compare our results with proximation and detail as low- and high-pass filter operations on the
[1]. However, our proposed system is capable of identifying any fault original signal, respectively.
in any analog circuit which can be represented by a unique signature. In applying wavelet analysis to sampled signals, one needs to per-
Comparison of our work with [1] in a later section clearly indicates form the downsampling operation after each level of decomposition to
the importance of preprocessing the output of an analog circuit in: 1) preserve the number of data points in the original signal. This simply
significantly reducing the size and, therefore, the training time of the means to include every other data point in the signal components (ap-
neural network and 2) improving the performance of the neural net- proximation and detail). Consequently, the number of data points in
work in fault diagnosis. the components at level j approximation or detail will be reduced by a
The material in this brief is arranged in the following order. In Sec- factor of two compared to the corresponding number of data points at
tions II and III, we briefly review our proposed preprocessing tech- level (j 0 1): This can be effectively used to extract optimal features
niques for analog-circuit fault diagnosis and cover the sample circuits for neural network training as explained later. The process of down-
and faults considered in this work. Section IV discusses the selection of sampling is also shown in Fig. 1.
wavelet coefficients as features to train and test the neural network, and Several families of wavelets are proven to be useful in signal and
Section V covers the results. We will present our conclusion in Section image processing. These families include Haar, Daubechies, Biorthog-
VI. onal, Coiflets, Symlets, Morlet, Mexican Hat, Meyer, etc. Each of these
wavelets have properties that make them suitable for certain applica-
tions [7]. For example, the wavelet Biorthogonal 3.7, which belongs
II. PROCESSING OF NEURAL NETWORK INPUTS to Biorthogonal family of wavelets, is popular in image processing,
A. Wavelet Transform as a Preprocessor and the Mexican Hat wavelet has been applied to rotation-invariant
pattern recognition [8]. The proper choice of the mother wavelet for
Discrete (dyatic) wavelet analysis refers to the decomposition preprocessing the analog circuit’s output(s) is crucial for optimal de-
of a signal into the so-called approximations and details. This is sign of the fault diagnostic system. In this work, we have used the first
accomplished using shifted and scaled versions of the so-called member of Daubechies family of wavelets, also known as Haar, shown
original (mother) wavelet as in Fig. 2 for the circuits under test. This choice was made after sev-
eral wavelet functions were examined as possible candidates for the
mother wavelet. After analyzing the wavelet coefficients generated by
x
( )= p1a x0b : (1)
these wavelet functions, it was determined that Haar gave the most dis-
a;b
a tinct features (wavelet coefficients) across fault classes, thus providing
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS—II: ANALOG AND DIGITAL SIGNAL PROCESSING, VOL. 47, NO. 2, FEBRUARY 2000 153

Fig. 2. The mother wavelet used in our study. This is the first member of the
Daubechies family of wavelets, also known as Haar. Fig. 3. A 25-kHz Sallen-Key bandpass filter used in our study.

optimal features for neural network training. This issue is discussed in which two or more input features may differ by several orders of
further in Section IV. magnitude. These large variations in feature sizes can dominate more
As mentioned before, the main advantage of using wavelet transform important but smaller trends in the data and should be removed through
is to reduce the number of inputs to the neural network. This is accom- normalization. Our studies have shown that normalizing feature vec-
plished by the downsampling operation performed in wavelet analysis, tors selected by PCA to have zero mean and unit standard deviation
which causes the number of wavelet coefficients to be reduced by a can make the neural network training phase much more efficient.
factor of two at each level of approximation or detail. Since this de- The neural network selected for this study is a multilayer feedfor-
composition is hierarchically organized, we can search for an appro- ward neural network trained by backpropagation [9]. Our work shows
priate collection of these coefficients that remain distinct across fault that through preprocessing of the output signals from an analog circuit,
classes. The selected coefficients are further preprocessed by PCA and we can select an optimal number of features (inputs) for the neural
normalization to prepare the inputs to the neural network. The coeffi- network. This consequently minimizes the size of the neural network,
cient selection process will be discussed later in Section IV. reducing its training time and improving its performance. The wavelet
coefficients not selected as features to train the neural network are dis-
B. PCA and Data Normalization carded since they are irrelevant to distinguishing among fault classes.
Principal component analysis is a preprocessing technique which can
significantly reduce the complexity of the neural networks employed III. SAMPLE CIRCUITS AND FAULTS
in fault classification problems. PCA achieves this goal by further re- The two circuits studied in our work are the same as those in [1] and
ducing the dimensionality of the input space after wavelet analysis are shown in Figs. 3 and 5. The first circuit is the Sallen-Key band-
while preserving as much of the relevant information as possible for pass filter [10]. The nominal values for the components which result
fault classification. In cases where input space is not high-dimensional, in a center frequency of 25 kHz are shown in the figure. The resistors
PCA can still be a very useful preprocessing technique for selecting op- and capacitors are assumed to have tolerances of 5% and 10%, respec-
timal features that minimize classification errors. In summary, the aim tively1. The primary motivation for selecting this filter and its associ-
of PCA is to map vectors x in a d-dimensional input space (x1 ; 1 1 1 ; xd ) ated faults described later in this section is to compare our results with
onto vectors z in an M -dimensional space (z1 ; 1 1 1 ; zM ) in such a way those in [1]. The same circuit and faults are considered in this reference
that loss of essential information and data variation needed for classifi- with the impulse response sampled and fed directly to a neural network
cation is minimized while keeping M < d: To apply PCA to a typical without any preprocessing. Comparison of the two case studies proves
input space vector x; we first write x as a linear combination of a set the importance of preprocessing in neural network based diagnostic
of d orthonormal vectors ui in the form systems which leads to neural networks having simpler architectures
d and improved performance.
x= i ui z (3) The impulse response of the circuit in Fig. 3, with R3; C 2; R2; and
i=1 C 1 varying within their tolerances, belong to the no-fault class (NF)

where i = uTi x Next, we retain only a subset


z : of the basis
M < d and are fed to the preprocessors for feature selection. When any of the
vectors ui and their associated i values to form
z four components is higher or lower than its nominal value by 50% with
M d the other three components varying within their tolerances, we obtain
x =0
i ui +
z i ui b (4) faulty impulse responses. These faulty impulse responses are similarly
i=1 i=M +1 fed to the preprocessors for feature selection and form the fault classes
where x is the approximation to x In (4), the remaining coefficients R3 *; R3 +; C 2 *; C 2 +; R2 *; R2 +; C 1 *; and C 1 +; where * and
0
:

zi for  + 1 are replaced by constants i Minimizing


i M b : + stand for high and low, respectively. For instance, R3 * fault class
N n n2 corresponds to R3 = 3k; with C 2; R2; and C 1 allowed to vary within
M = 12
E kx 0 x k (5)
their tolerances. This procedure allows us to generate training and test
n=1 patterns at the output of the preprocessors for training and testing the
neural network. The impulse response of the filter can be very well
which is the sum of the squares of the errors over the whole data set approximated by the output generated from a narrow pulse whose width
(n = 1; N ); gives the vector z with components zi (i = 1; M ); which T is much smaller than the inverse of the filter’s bandwidth. For all our
approximate x in the M -dimensional input space. In practice, M is SPICE simulations, we have taken the filter input to be a single pulse
determined by principal component analysis such that a pre-specified of height 5 V and duration of 10 µs. Our simulation results indicate
percentage of the total variation in the data set is preserved.
1The robustness of the neural network based fault diagnostic system to the
The third preprocessing stage is data normalization which follows
tolerance of nonfaulty components is automatically verified during a successful
PCA for the purpose of enhancing the features in a data set. Data nor- training phase. This phase requires the neural network to place output voltages
malization is a linear scaling of the input features to avoid large dy- corresponding to the given tolerance ranges in the same fault class and achieve
namic ranges in one or more dimensions. There are many applications a prespecified error goal.
154 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS—II: ANALOG AND DIGITAL SIGNAL PROCESSING, VOL. 47, NO. 2, FEBRUARY 2000

Fig. 4. Linear plot (value versus index) of the first element of levels 1–5 approximation coefficients for the circuit shown in Fig. 3. Fault classes in each plot are
in the order 1 1 2 2 2 2 3 and 3 and are separated from each other by vertical lines. Within a fault class,
50 typical values are shown.

that the output generated by such a pulse is very adequate to classify The reason these faults belong to the same ambiguity group is because
all faults considered here. they produce very similar outputs, as discussed in Section V.
The second circuit studied in our work and [1] is more complicated
and is shown in Fig. 5. This is a two-stage four-op-amp biquad low-pass IV. WAVELET COEFFICIENT (FEATURE) SELECTION
filter [10]. The work in [1] uses two ambiguity groups of seven and
eight faulty components. The first group contains C 1 *; C 3 *; C 4 *;
As indicated before, we have used Haar wavelet function in this work
R16 *; R19 *; R21 +; and R22 *; and the second group consists
since it gives the most distinct features across fault classes. This be-
of C 2 *; R17 +; R3 *; R4 +; R6 *; R7 +; R8 *; and R9 *.
comes clear during the training phase, when the network must meet
a reasonable error goal to warrant good generalization. Among many
The faulty component values used in our work and [1] are shown in
Table I where * and + imply significantly higher and lower than nom-
wavelets we examined, only Haar function achieved our prespecified
mean-square error goal of 0.01. The wavelet properties that can pro-
inal values. These two ambiguity groups are defined in [1] based on
vide insight into the appropriateness of this function are support and
the degree of similarity between circuit outputs belonging to different
regularity. Support is a measure of the duration of a wavelet in time
fault classes. The measure used in this reference to obtain the degree
domain and regularity is related to its ability to correlate to smooth
of similarity is the K1 criterion defined by
functions [11]. Since Haar function has a compact support and a regu-
larity of zero because of its discontinuous nature, it is very well suited
Rxy to extract features from signals characterized by short durations and
K1 =
R xx + Ryy 0 xy
R
(6)
swift variations. When an input pulse is applied to our circuits, it is the
localized behavior of the output signal at its onset that carries the dis-
where tinct features required to classify faults. This localized behavior of the
T output signal at its rupture characterized by short duration then makes
1 Haar wavelet appropriate for this application.2 The general guidelines
R xx = x(t)x(t) dt (7)
T 0 for selecting wavelet coefficients as features to train the neural network
are as follows:
and
1) Depending on the nature of the signal, one needs to go up to
1
T sufficiently high levels of approximation and detail to expose the
R xy = x(t)y (t) dt: (8) low- and high-frequency features of the signal, respectively. One
T 0
can then use principal component analysis to further reduce the
In this equation, x(t) and y (t) are the signals whose correlation is 2To demonstrate the significance of the localized behavior of the outputs in
under investigation and T is the period. If this measure indicates high this application, we have examined these signals, which typically last 1 ms. Ex-
correlation between outputs, the associated faults are placed in the same tracting features from the first 0.2 ms of the output signals is sufficient to train
ambiguity group. Through our proposed preprocessing techniques, we the neural network and achieve a mean square error goal of 0.01. Training the
neural network on features extracted from the last 0.8 ms of the output signals
show that it is possible to select an optimal number of features that leads to a mean square error which is several order of magnitudes higher. This
leads to successful classification of individual faults presented in Table proves that the onset of the output signal has the necessary information to dis-
I except for one ambiguity group containing R6 *; R7 +; and R9 * : tinguish among fault classes which is best analyzed by the Haar function.
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS—II: ANALOG AND DIGITAL SIGNAL PROCESSING, VOL. 47, NO. 2, FEBRUARY 2000 155

Fig. 5. The two-stage four-op-amp biquad low-pass filter used in our study. All resistors are in ohms.

number of features and/or enhance their distinctiveness across has 49 inputs, 10 first-layer, and 10 second-layer neurons, resulting in
fault classes. a total adjustable parameters of about 700. During the training phase,
2) The approximation and detail coefficients reflect the low- and an error function of these parameters must be minimized to obtain the
high-frequency contents of a signal, respectively. The low-fre- optimal weight and bias values. Their trained network was able to prop-
quency contents usually give a signal its basic structure, while erly classify 95% of the test patterns. Our backpropagation neural net-
the high-frequency contents provide its details. As a result, the work has two layers with 5 inputs, 7 neurons in layer 1, and 1 neuron
main features of a signal are usually captured in the approxima- in the output layer. The total number of adjustable parameters in our
tion coefficients. network is about 45, with the reduction in the number of weights and
Using these guidelines, we have selected the first coefficient of approx- biases directly translating to shorter training time and better perfor-
imation levels 1–5, associated with the impulse response of the filters as mance. Our trained network is capable of 100% correct classification
our features. Training the neural network on these features meets our of our test data.
prespecified error goal of 0.01, implying that they are distinct across The real advantage of preprocessing becomes evident when applied
fault classes. It is possible to demonstrate graphically the role these to the more complex low-pass filter shown in Fig. 4. As mentioned be-
features play in distinguishing among fault classes. Consider the NF fore, the work in [1] assigns all faults to one of two ambiguity groups.
and eight faulty classes associated with the sample circuit in Fig. 3. The size of the neural network is not specified in their work and they
For each of the nine classes, 50 impulse responses are generated by achieve a 100% correct classification of the test data as belonging to
varying components within their tolerances as described in Section III. one of the two ambiguity groups. We have successfully classified all the
The selected approximation coefficients associated with these impulse faults in Table I except three, (R6 *; R7 +; R9 *); which are placed in
responses are shown in Fig. 4. In this figure, each plot corresponds to one ambiguity group. The backpropagation neural network trained for
a single coefficient with the x-axis ranging from 1 to 450 to cover nine this problem has 5 inputs, 20 neurons in hidden layer 1, and 1 neuron
intervals of size 50, separated by vertical lines, corresponding to each in the output layer. This neural network correctly classifies 99.3% of
class. Fault classes in each plot are presented in the order C 1 *; C 1 +; the test data. Due to the effective preprocessing of the low-pass filter’s
C 2 *; C 2 +; N F; R2 *; R2 +; R3 *; and R3 + : We can examine output, the neural network architecture used for this complicated cir-
Fig. 4 to determine how fault classes are distinguished by features. For cuit is very simple, implying fast and efficient training and superior
instance, feature 1 can separate N F; R3 *; and R3 +; or N F and performance.3 As mentioned before, the reason for R6 *; R7 +; and
R2 * can be distinguished by features 2 or 3. The effectiveness of R9 * belonging to the same ambiguity group is that they create very

these five features to classify the faults associated with the sample cir- similar outputs. This is evident from an analysis of the low-pass filter,
cuits will be discussed in the next section. It is important to note that which gives
feature selection is a critical and intricate task in analog fault diagnosis
dvo1 R7
or any other neural network based system. However, this task needs to = v1 : (9)
dt R6 R9 C2
be carried out only once in a given application.
For these three fault classes, the expression (9) remains relatively un-
V. RESULTS changed for the faulty component values shown in Table I. As a result,
3It is important to note that the work in [1] is a black-box approach to fault di-
In this section, we compare the size and performance of our neural
networks with [1] to show the significance of the proposed prepro- agnosis and does not analyze the circuit characteristics, except for some simple
signal similarity metric. This is intended to move the test/diagnosis objective
cessing techniques. To perform a diagnosis of the faults described in close to the goal of a built-in self-test module. However, in light of the advan-
Section III for the Sallen-Key bandpass filter, the work presented in [1] tages of the proposed preprocessing techniques, it is beneficial to add a prepro-
requires a three-layer backpropagation neural network. This network cessing unit to this module.
156 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS—II: ANALOG AND DIGITAL SIGNAL PROCESSING, VOL. 47, NO. 2, FEBRUARY 2000

TABLE I REFERENCES
FAULT CLASSES USED FOR THE TWO-STAGE
FOUR-OP-AMP BIQUAD CIRCUIT. THE NOMINAL AND FAULTY [1] R. Spina and S. Upadhyaya, “Linear circuit fault diagnosis using neuro-
VALUES ARE ALSO SPECIFIED morphic analyzers,” IEEE Trans. Circuits Syst. II, vol. 44, pp. 188–196,
1997.
[2] H. Spence, “Automatic analog fault simulation,” in Proc. AU-
TOTESTCON’96 Conf., pp. 17–22.
[3] R.-W. Liu, Testing and Diagnosis of Analog Circuits and Sys-
tems. New York: Van Nostrand, 1991.
[4] Selected Papers on Analog Fault Diagnosis. New York: IEEE Press,
1987.
[5] J. W. Bandler and A. Salama, “Fault diagnosis of analog circuits,” in
Proc. IEEE, 1985, pp. 1279–1325.
[6] M. Catelani and M. Gori, “On the application of neural networks to
fault diagnosis of electronic analog circuits,” Measurement, vol. 17, pp.
73–80, 1996.
[7] G. Strang and T. Nguyen, Wavelet and Filter Banks. Cambridge, MA:
Wellesley-Cambridge Press, 1996.
[8] H. W. Lee, Wavelet Gri-Minace Filter for Rotation-Invariant Pattern
Recognition. Philadelphia, PA: SPIE, 1994, vol. 2762, pp. 343–352.
[9] M. H. Hassoun, Fundamentals of Artificial Neural
Networks. Cambridge, MA: MIT Press, 1995.
[10] M. V. Valkenburg, Analog Filter Design. New York: Oxford Univ.
Press, 1982.
[11] B. B. Hubbard, The World According to Wavelets. Boston, MA: Peters,
1996.
[12] G. J. Hamink, B. W. Meijer, and H. G. Kerkhoff, “Testability analysis
of analog systems,” IEEE Trans. Computer-Aided Design, vol. 9, pp.
573–583, June 1990.
these three faults must be placed in the same ambiguity group if the cir-
[13] C. M. Bishop, Neural Networks for Pattern Recognition. New York:
cuit output is the only measure to determine fault classes. Training the Oxford Univ. Press, 1995.
neural network on more than one node voltage or using testability anal-
ysis as a preprocessor [12] can further resolve the ambiguity groups.
For instance, if the training data for the lowpass filter includes features
associated with the v2 node voltage, our neural network can classify
R9 * fault correctly and the ambiguity group reduces to R6 * and

R7 + : Our results clearly indicate that through appropriate prepro-


Cascaded Parallel Oversampling Sigma–Delta Modulators
cessing of an analog circuit output, one can train a neural network to
correctly diagnose all faults unless the circuit’s outputs are similar for Xuesheng Wang, Wei Qin, and Xieting Ling
some fault classes.

Abstract—Based on the well-known time-interleaved modulator (TIM), a


new cascade-parallel architecture of oversampling sigma–delta analog-to-
VI. CONCLUSION digital converters is proposed. While retaining the speed advantage of TIM,
the new architecture gives a general method to effectively suppress the in-
fluence of circuit nonidealities, especially coefficient mismatches, on the
We have applied backpropagation neural networks with wavelet de- converter’s resolution. Such influence is a serious problem in the practical
composition, PCA, and data normalization as preprocessors to fault di- realization of TIM. Simulation results of examples of both TIM and the new
agnosis of analog circuits. Our study indicates that the proposed pre- architecture are given for comparison. In addition to its improved perfor-
processing techniques have a significant impact on analog fault diag- mance, the new architecture turns out to be quite simple. Therefore it can
be a practical approach to extend the use of sigma–delta analog-to-digital
nosis due to the selection of an optimal number of relevant features. conversion to high-speed applications.
This leads to neural network architectures with minimal size that can
be trained efficiently and carry out fault diagnosis with a high degree Index Terms—Cascade, converters, parallel, TIM.
of accuracy. For complex analog circuits, our approach leads to neural
networks that can identify individual faulty components unless these I. INTRODUCTION
faults give similar circuit outputs. In such a case, the neural network can
identify the ambiguity group containing the fault classes with similar Because of their outstanding linearity, oversampling converters have
outputs. Further resolution of ambiguity groups is possible by addition become a popular technique for data conversion [2]. However, due to
of features from more node voltages to train the neural network. Since the nature of oversampling, these converters are much slower than their
our analog fault-diagnostic system is based on neural networks, its per- Nyquist-rate counterparts. Hence, the applications of sigma–delta mod-
formance depends on how well the selected features can distinguish ulators are usually restricted to low-speed high-linearity applications
among fault classes. As a result, choosing proper wavelet function and such as digital audio. Emerging needs have forced designers to seek
wavelet coefficients is critical to the system performance. Since one highly linear converters with broader input bandwidths.
can not make these choices with absolute certainty a priori, any ad-
ditional analysis to ensure the distinctiveness of the selected features
across fault classes is beneficial [13]. The test that will eventually deter- Manuscript received January 1999; revised October 1999. This paper was
recommended by Associate Editor H. Tanimoto.
mine the appropriateness of the selected wavelet function and features The authors are with the Department of Electrical Engineering, Fudan Uni-
is the training phase which must meet a reasonable error goal leading versity, Shanghai, China.
to a satisfactory performance of the neural network. Publisher Item Identifier S 1057-7130(00)01463-4.

1057–7130/00$10.00 © 2000 IEEE

You might also like