You are on page 1of 62

CHAPTER 16

SIGNAL PROCESSING IN ROTATING MACHINERIES

In previous chapter, we studied various kinds of transducers, and other measuring and analyses
equipments used in the vibration and acoustics. These transducers produce the signals in the form of
charge or voltage. Through amplifiers these signals are fed to the data acquisition system to digitize
these signals into a computer. In the present chapter, the main focus would be to study methods
involved in the processing these signals in various required form in time and frequency domains (fast
Fourier Transform function, FFT; etc.) or time-frequency domain (wavelet transform, WT; etc.).
Various graphical forms of display of these measurements are presented. Error involved in the
measurement (mechanical, electrical or digital) itself would be addressed. These include error due to
the aliasing effect, error in digitizing, leakage error, tuning of sampling and signal periodicity, effects
due to windowing, etc. Fourier series, Fourier transform, discrete Fourier transform, FFT and
complex-FFT would be covered in some detail. The uncertainty in the estimated parameters due to
error in measurements would be addressed. In the subsequent chapter, various fault conditions and its
effects on the vibration signature are looked in.
16.1 Display of Vibration Measurements
Vibration signal collected has to be measured and displayed. The quantities that have to be measured
are usually vibration amplitude, its frequency and its phase. The history of the vibration measurement
in the form of digital display (through vibration meter) gives only the trend of vibration level (peakto-peak, rms, etc.), which might give only indication about the condition of the machine whether
machine is in good or bad condition. However, the cause of the increased vibration level at the
component level is difficult to assign with such lone measurement of vibration level by vibration
meter. To investigate the causes of vibration in component level, one needs to investigate the
relationship between the vibration signal frequencies and the rotational speed of the shaft. In
principle, every fault has a unique frequency (or frequencies) to contribute in the overall vibration
signal, especially in rotating machinery different faults at the component level might contribute
vibration frequencies in integer multiples (or divisions) of rotational frequency of component speeds.
This can be done by the spectrum analysis with the help of FFT-analyzer (fast Fourier transformation)
equipments. Spectrum analysers have various convenient functions, such as FFT, tracking analysis
(Bode plot, Nyquist plot), waterfall diagram, Campbell diagram, etc.
A tracking filter is a device that accepts two input signals, one being the vibration signal under
consideration and the other being a phase reference signal. In the tracking analysis, dynamic

946
characteristics of a rotating machine are investigated by measuring vibrations for the run-up or
rundown of the machinery, which is a practical way of getting dynamic behaviour of the machinery at
various operational speeds. The tracking of the vibration signal is done normally at 1 rotational
speed of the rotor. The tracking filter removes from the vibration signal any components, which are
not of the same frequency as the reference signal. The amplitude and phase (relative to the reference
signal) of the remaining signal are then displayed, usually in digital form. The output from a tracking
filter can be used to construct the Bode and Nyquist diagrams used, for example, during balancing
operations. This can be done several ways for example (i) by manually tuning the narrow bandpass
filter to the required central frequency (1 or integer multiples of rotational speed of rotors) and
measuring the vibration level by vibration meter, (ii) tuning the central frequency as in method (i),
however, displaying the signal in the oscilloscope to take required measurement, and (iii) measuring
the vibration signal continuously for the total run-up or rundown period and processing the signal
afterwards in computer. Apart from the amplitude of vibration often it is required to measure the
phase of the vibration signal with respect to some fixed reference on the rotating component, which is
called the reference signal. Phase meter can be used for method (i) which requires two signals, i.e.,
first the vibration signal and the other is the reference signal. For method (ii) both the signals could be
displayed on the same screen and relative phase can be measured. In method (iii) vibration signals and
the reference signal have to recaptured at various small time period and amplitude and phase of the
vibration signal can be measured by displaying it on the computer screen or by advance signal
processing software.

16.1 Bode plot (the variation of the amplitude and


phase with the rotational frequency of rotor)

Fig. 16.2 Nyquist plot

947
The vibration amplitude and the phase so obtained, which is called the frequency response function,
can be displayed as the Bode and Nyquist plots as shown in Figures 16.1 and 16.2 (the Bode plot as
shown is for multi-DOF system whereas the Nyquist plot is plot of one of the resonance peak or for
single-DOF system). The amplitude of the Bode plots should be plotted in semi-logarithmic scale
when the peak at the resonance is very sharp with very high magnitudes. The Nyquist plot to be
considered as blown view of the resonance peak and this plot is very useful for obtaining the modal
damping (Ewins, 1984). In the Nyquist plot, it is shown that both the amplitude and phase information
of the Bode plot is merged in a single Nyquist plot. The frequency information is indicated along the
circumference of the circle of the Nyquist plot. It can easily be seen that the Nyquist plot is a complex
plane with the horizontal axis as real axis and the vertical axis as the imaginary axis. This will be
more clear when we see the Bode plot information can also be plotted as the real and imaginary part
of the signal with respect to the frequency (i.e., Xe j = X cos + jsin = X r + jX i ; where X is the
amplitude, is the phase, X r is the real part and X i is the imaginary part).

Figure 16.3 Spectrum diagram (FFT of a capture signal)

Figure 16.4 Waterfall diagram (FFT of captured signals at different speeds)

948
Both the diagrams (Bode and Nyquist plots) should be plotted using readings, which have been
nulled at slow speed and between critical speeds, thereby removing the effects of runout and the
effects of vibration modes which do not relate to the critical speed under consideration. Depending
upon the vibration parameters, i.e., the displacement, velocity and acceleration, the FRF (frequency
response function, i.e, the ratio of the response to the force) is called respectively the receptance,
mobility and inertance. The inverse of FRF corresponding to the displacement, velocity and
acceleration measurements are called respectively the dynamic stiffness, the mechanical impedance
and the apparent mass.
The spectrum diagram is the plot of captured unfiltered vibration signal in frequency domain (or may
be filtered with the broad bandpass filter to remove the unnecessary noise). It is obtained by
performing the FFT of the captured signal and one such spectrum diagram is shown in Figure 16.3.
The spectrum analyser is used to separate out the incoming vibration signal into all of frequencies
from which the total signal is composed. The amplitude and phase, relative to some reference signal,
of all the frequency components is displayed in the form of a graph of amplitude (or phase) against
frequency. The data is thus said to be displayed in the frequency domain.

Figure 16.5 Campbell diagram


Many spectrum analyzer have the facility to plot several such graphs in cascade or waterfall
diagram, as shown in Figure 16.4. Such diagrams help to determine the relationship between the
vibration signal and machine running speed, which in turn help in the monitoring of the machine
condition. A waterfall diagram is a 3-dimensional plot of spectra at various speeds. Vibration signals
are captures at different speeds and FFT of the signals are cascade as shown in Figure 16.4. Another
version of the waterfall diagram in 2-dimension is shown in Figure 16.5, where the amplitudes of the
vibrations are depicted by a circle with radius corresponding to the amplitude of vibration at particular
frequency (i.e., 1 or integer multiples/divisions). The axis is chosen similar to the Campbell diagram,
i.e., spin speed of the rotor as abscissa and frequency of whirl as the ordinate.

949

(a) The shaft and transducers

(b) Measured and reference signals


Figure 16.6 The measurement of the phase in a rotor
In rotor applications where phase is to be measured (as in constructing the Bode and Nyquist plots), a
reference signal indicating a particular location on the shaft will be required in addition to the
vibration signal which is to be measured as shown in Figure 16.6. This key phasor or reference signal
might be obtained by using (i) a proximity transducer to detect the passing of a key way or a bolt head
on the shaft or (ii) an optical transducer detecting the passing of photo-sensitive tape adhered to the
shaft or alternatively some other signal which produces a well-defined peak can be used. The key
phasor signal is used in conjunction with the vibration signal; the phase angle recorded being that
between the peaks of the two signals (see Figure 16.6(b)). An oscilloscope provides an immediate
visual representation of the vibrational signal, enabling the engineer to observed the form of the signal
(sine wave or otherwise), its frequency and its amplitude. This is known as time domain data
representation.
Most oscilloscope provide x and y facility which allows two vibration signal from, say two
proximity transducers maintained perpendicular to each other, to be input to the oscilloscope. These
signals can be displaced in the form of a shaft whirl orbit instead of the normal time-base

950
representation. The vibration of the rotor is a whirling motion and therefore not only the frequencies
but also the directions of the whirling motions are important enough to pursue their causes. Some
oscilloscopes enable a phase-indicating pulse of extra-bright beam to be displaced on the orbit once
per revolution of the shaft, which indicates the direction of whirl. The shape of the orbit, or Lissajous
figure, can itself be a useful tool in monitoring machine health. A typical whirl orbit for a machine
subject to a small amount of unbalance is shown in Figure 16.7 (as displayed on an oscilloscope).

Figure 16.7 Lissajous plot


The orbit plot can be seen clearly for simple rotor vibration when a single frequency is present in the
measured signals. However, using manual tracking filter the orbit plot can be obtained for 1 or
integer multiples/divisions of the rotational frequency of the rotor, which is a tedious process. In
frequency domain , however, since the usual fast Fourier transform (FFT) theory gives information
about magnitudes of frequencies and phases only, we cannot know the whirl direction using the
conventional FFT-analyser. For this purpose, Ishida (1997) and Lee (2000) proposed a signal
processing method where the whirling plane of a rotor is overlapped to the complex plane. This
method is called the complex-FFT (or directional-FFT) method, enables us to know the directions of
whirling motion besides the magnitudes of the frequencies. They also used this method to extract a
component form non-stationary time histories obtained numeric simulations and experimented data
and depicted the amplitude variation of the component. We will discuss fundamental ideas necessary
to understand the signal processing by computer. In addition, applications of the complex-FFT
method in studying stationary and non-stationary vibrations are explained.
Once the vibration signals have been collected and measured, they are used to judge whether or not
the machine in question is operating properly. These judgments are made on the basis of whether
there are unusual features of vibration signal which are not normally present, or which may usually be
present under particular fault conditions. However, to generate such plots and use various functions
correctly, we must have some background knowledge of signal processing. Moreover, if we have to
construct a specific data analysis system that fits our research for example for system parameter
estimation (i.e., the crack parameter, bearing and seal dynamic parameters, residual unbalance,

951
misalignment, etc.), we must have sufficient understanding of the fundamental of the signal
processing.
16.2 Accuracy of Vibration Measurements
A basic measurement system consists of a transducer with a conditioner (which senses the signal, and
converts to a measurable quantity like voltage or charge with some amplification), an analog-todigital converter, and a processing and display unit as shown in Figure 16.8. We shall focus our
attention in error involved in transforming the analog signal to the digital signal.

Transducer and
conditioner
(signal detection)

Interface
(A/D-transform)

Personal Computer
(Processing and display)

Figure 16.8 A basic measurement system


Vibration signal from the vibrating structure is detected and converted into a measurable physical
quantity by transducers. In a rotating machine, rotor displacements in two orthogonal directions and a
rotating speed are detected as voltage variations. The output signal x(t) from the transducer is an
analog signal that is continuous with time (it is similar to a theoretical function that is defined for all
possible values of its variables) as shown in Figure 16.9(a). However, the signal is discretised when it
is acquired by computer through an A/D interface as shown in Figure 16.9(b), where the solid line
represent the actual signal and circles represent the digitized signal data at grid nodes. It can be seen
that depending upon the resolution of the instrument the accuracy of the acquired data will depend,
finer the grid more the accuracy and vice a versa, since the digitized data will be aquired only at the
grid points. A digital signal is discretised in both time and magnitude. This digital signal is a series of
discrete data {x n } obtained by measuring (called sampling) an analog signal instantly at every time
interval t and is given as x n = x(nt ) , where n is an integer. This interval t is called a sampling
interval, which is generally constant. Discretization in magnitude is called quantization, and the
magnitude is represented by binary numbers (unit: bit. A bit (short for binary digit) is the smallest unit
of data in a computer. A bit has a single binary value, either 0 or 1. Although computers usually
provide instructions that can test and manipulate bits, they generally are designed to store data and
execute instructions in bit multiples called bytes. In most computer systems, there are eight bits in a
byte). Binary describes a numbering scheme in which there are only two possible values for each
digit: 0 and 1. In binary numbers the digits'weight increases by powers of 2, rather than by powers of
10 as in the more familiar decimal numbers. In a digital numeral, the digit furthest to the right is the
"ones" digit; the next digit to the left is the "twos" digit; next comes the "fours" digit, then the "eights"
digit, then the "16s" digit, then the "32s" digit, and so on. The decimal equivalent of a binary number

952
can be found by summing all the digits. For example, the binary 10101 is equivalent to the decimal 1
+ 0 + 4 + 0 +16 = 21.

x(t )

xn

t
(a) Analog signal

(b) Digitised signal

Figure 16.9 The analog and digital signals


Now we shall highlight some other errors which may be involved in measuring the vibration signal.
When an analog signal x(t) is changed into a sequence of digital data {x n } (n = 0, 1, 2, , N) a
virtual (or imaginary) wave is obtained if a fast signal is sampled slowly. For example, when a signal
illustrated by the full line is sampled as shown in Figure 16.10, a virtual signal wave illustrated by the
dashed line appears, although it is not contained in the original signal.

Figure 16.10 Aliasing effect on a fast signal while slow sampling

953

=10 Hz

x(t)

0.5

-0.5

-1

0.2

0.4

(a)

0.6

0.8

Time, s
1

f s = 2 Hz

x(t)

0.5

- 0.5

-1
0

0.1

0.2

(b)

0.3

0.4

0.5

Time, s

f s = 17 Hz

x(t)

0.5

-0.5

-1
0

(c)

0.1

0.2

0.3

0.4

0.5

Time, s

0.6

0.7

0.8

0.9

954

f s = 20 Hz

x(t)

0.5

-0.5

-1
0

0.2

(d)

0.4

0.6

0.8

Time, s

f s = 24 Hz

x(t)

0.5

-0.5

-1
0

0.2

0.4

(e)

0.6

0.8

Time, s

f s = 30 Hz

x(t)

0.5

-0.5

-1
0

(f)

0.2

0.4

0.6

Time, s

0.8

955

f s = 40 Hz

x(t)

0.5

- 0.5

-1
0

0.2

0.4

0.6

0.8

Time, s
f s = 50 Hz

x(t)

0.5

-0.5

-1
0

0.2

0.4

0.6

0.8

Time, s
Figure 16.11 Illustration of aliasing effect and sampling frequency in a sinusoidal signal
This phenomenon is called the aliasing. This phenomenon could be seen in old movies where the
wheel of a horse-cart appears rotating in the opposite direction even when the cart moves forward.
This is due to the fact that the picture frame recorded per second by the camera is much slower than
the rotational speed of the wheel. Another example could be that when we look an accelerating fan it
appears to us in between as if it is rotating in the backward direction (or in the forward direction with
very slow speed). It is again due to the fact that the human sight can capture pictures at a certain rate
and when the fan speed is just slightly higher (or lower) than the rate of picture frame captured by the
sight, then the fan appears as if it is rotating in opposite direction (or at different speed then it is
actually rotating). It is more obvious when we put stroboscope light on a rotating shaft. In stroboscope
the light flashes (switches on and off) at a particular rate, which can be adjustable. When the flashing
rate is same as the shaft speed (or integer multiples) the shaft appears as if it is stationary. It due to the
fact that whenever light flashes the shaft occupies the same orientation and if we see a mark (or

956
keyway) on the shaft, it will appear stationary. A slight decease in the flashing rate makes the shaft to
appear as rotating slowly in the same direction as the actual direction (when flashing rate is slightly
lower than the shaft speed the mark on the shaft will shift in forward direction in the subsequent
flashes and hence the shaft will appear as it is rotating in forward direction with slower speed i.e. with
a speed equal to the difference between the actual shaft speed and flashing rate), however, a slight
increase in the flashing rate makes the shaft to appear as if rotating slowing in backward direction as
compared to the actual direction of rotation. It is obvious that we must sample with a smaller
sampling interval as the signal frequency increases. This suggests that aliasing effects not only
changes the amplitude and frequency but also the whirl direction of a rotor. We can determine
whether or not we have this aliasing by following the sampling theorem. It says: when a signal is
composed of the components whose frequencies are all smaller than f c , we must sample it with a
frequencies higher than 2 f c for the sake of not loosing the original signals information. The
frequency 2 f c is called the Nyquist frequency. For example, if a sine wave with period T is sampled
whenever x (t ) = 0 , that is, with sampling interval T/2, we have xn = 0 (i.e., a signal with a constant
amplitude of zero). Moreover, if it is sampled whenever x(t ) = A (where A is the amplitude), that is,
with sampling interval T/2, we have xn = A (i.e., a constant amplitude signal).Therefore, two
samplings in a period are clearly insufficient. However, this theorem teaches us that digital data with
more than two points during one period can express the original signal correctly. For example, if we
sample the signal having components of 1, 2 and 7 kHz with a sampling frequency of 10 kHz. Then 1
and 2 kHz signal will be measured without aliasing effect since the Nyquist frequency is 10/2 = 5 kHz
and is more than these vibration signal frequencies. However, we have an imaginary spectrum of 3
kHz (10 kHz - 7 kHz = 3 kHz), which does not exist practically. But, if we sample it with a frequency
of more than 14 kHz ( 2 7 kHz), such an aliasing problem does not occur. In practical
measurements, we do not commonly determine the sampling frequency by trial measurement. Instead,
we use a low-pass filter to eliminate the unnecessary high-frequency components in the signal and
sample with the frequency higher than twice the cutoff frequency. By such a procedure, we can
prevent aliasing.

Example 16.1 If we sample the signal having components of 120, 200, 460, 700, 800 and 900 Hz
with a sampling frequency of 1000 Hz. Whether aliasing effect will be present in the measurement?
What are the frequency that will appear in the captured signal?
Solution: The sampling frequency is 1000 Hz, hence the Nyquist frequency will be 1000/2 = 500 Hz.
Hence, we can be able to measure frequency below 500 Hz accurately that means frequency 120, 200
and 460 Hz will be measured accurately (Figs. 16.12-16.14). However, frequency 700 Hz will appear

957
as 1000-700 = 300 Hz (Fig. 16.15) and frequency 800 Hz will also appear as 1000-800 = 200 Hz (Fig.
16.16). Frequency 900 Hz will appear as 1000-900 = 100 Hz signal (Fig. 16.17). So the captured
signal will contain erroneous high amplitude of 200 Hz signal with an additional frequency of 100 Hz
which is actually not present at all in the actual signal.
8

0.5

FFT[x(t)]

x(t)

0
-0.5
-1
0

0.02

Time, s

FFT[x(t)]

x(t)

100

200

300

Frequency, Hz

600

0.5
0
-0.5

400

200

-1
0

0.02

0.04

x 10

15

FFT[x(t)]

0.5
0
-0.5
-1
0

100

200

300

Time, s
Frequency, Hz
(b) Sampled signal
Fig. 16.12 Signal of 120 Hz sampled with sampling frequency 1000 Hz

x(t)

(a) Original signal

0.02

Time, s

10
5
0

0.04

(a) Original signal

200

400

600

200

400

600

Frequency, Hz

600

FFT[x(t)]

0.5

x(t)

0.04

x 10

- 0.5
-1
0

0.02

0.04

400

200

Frequency, Hz
Time, s
(b) Sampled signal
Fig. 16.13 Signal of 200 Hz sampled with sampling frequency1000 Hz

958

FFT[x(t)]

x(t)

0.5
0
-0.5
-1
0

0.005

0.01

Time, s

0.015

500

1000

Frequency, Hz

600

FFT[x(t)]

x(t)

(a) Original signal

0.5
0
-0.5

400

200

-1
0

0.005

0.01

0.015

0.02

200

400

Frequency, Hz
Time, s
(b) Sampled signal
Fig. 16.14 Signal of 460 Hz sampled with sampling frequency1000 Hz

0.5

FFT[x(t)]

x(t)

0.02

-0.5
-1
0

0.005

0.01

Time, s

0.015

x 10

2
1
0

0.02

(a) Original signal

500

1000

1500

Frequency, Hz

600

FFT[x(t)]

0.5

x(t)

x 10

0
-0.5
-1
0

0.005

0.01

0.015

0.02

400

200

200

400

600

Frequency, Hz
Time, s
(b) Sampled signal
Fig. 16.15 Signal of 700 Hz sampled with sampling frequency1000 Hz

959

FFT[x(t)]

x(t)

0.5
0
-0.5
-1
0

0.005

0.01

Time, s

0.015

1000

1500

200

400

600

Frequency, Hz

600

FFT[x(t)]

0
-0.5

400

200

-1
0

0.02

0.04

Frequency, Hz
Time, s
(b) Sampled signal
Fig. 16.16 Signal of 800 Hz sampled with sampling frequency1000 Hz

x 10

1
0.5

FFT[x(t)]

x(t)

500

(a) Original signal

0.5

-0.5
-1
0

0.005

0.01

Time, s

0.015

500

1000

1500

Frequency, Hz

(a) Original signal

2000

600

FFT[x(t)]

0.5

x(t)

0.02

x(t)

x 10

0
-0.5
-1
0

0.02

0.04

400

200

100

200

300

400

Time, s
Frequency, Hz
(b) Sampled signal
Fig 16.17 Signal of 900 Hz sampled with sampling frequency1000 Hz

500

960
Example 16.2 A shaft is rotating at 30 Hz, what will be the speed of shaft which will appear to the
observer when the stroboscope flashing frequency is (i) 30 Hz (ii) 60 Hz (iii) 25 Hz (iv) 34 Hz (v) 54
Hz, (vi) 76 Hz and (vii) 600 Hz (viii) 15 Hz and (ix) 45 Hz (x) 13 Hz and (xi) 18 Hz. Give a plot of
such variation.
Solution: The following speed of the shaft will be observed (i) stationary (ii) stationary (iii) 25 30 =
- 5 Hz in the opposite direction as the actual shaft rotation (iv) 34 30 = 4 Hz in the same direction
the actual shaft rotation (v) 54 -60 = - 6 Hz in the opposite direction as the actual shaft rotation (vi) 76
60 = 16 Hz or 76- 90 = - 14 Hz; we will see 14 Hz in the opposite direction as the actual shaft
rotation (vii) stationary (viii) 15 -30 = - 15 Hz, since it is half the frequency of the shaft if we mark
two different colours on the diagonally opposite on the shaft surface we will see them in alternatively,
however, shaft will appear as rotating with the half actual speed of the shaft. (ix) 45-30 = 15 Hz, in
this also case if we mark two different colours on the diagonally opposite we will see them in
alternatively, however, shaft will appear as rotating with the half actual speed of the shaft (x) 13- 30 =
- 17 Hz in the opposite direction as the actual shaft rotation (xi) 18 -30 = -12 Hz, in the opposite
direction as the actual shaft rotation.
In general if f s is the shaft frequency and f sb is the frequency of the stroboscope then
(a) the Nyquist frequency would be 2 f s
(b) when f sb < f s then the shaft will appear as if it is rotating at frequency of f sb f s and the virtual
rotation of the shaft will be in the opposite to the actual rotation. However, when f sb = 0.5 f s if we
mark two different colours on the diagonally opposite we will see them in alternatively and the shaft
will appear as rotating with the half of actual speed of the shaft.
(c) when f sb = f s the shaft will appear as stationary.
(d) when f s < f sb < 1.5 f s the shaft will appear to rotate at frequency f sb f s in the same direction as
the actual.
(e) when f sb = 1.5 f s if we mark two different colours on the diagonally opposite we will see them in
alternatively and the shaft will appear as rotating with the half actual speed of the shaft.
(f) when 1.5 f s < f sb < 2 f s the shaft will appear to rotate at frequency f sb 2 f s in the opposite
direction as the actual.
(g) when f sb = 2 f s the shaft will again appear as stationary.
(h) In general, when f sb = nf s ( n = 1, 2,3,

) the shaft will appear as stationary and f sb = 0.5nf s if we

mark two different colours on the diagonally opposite we will see them in alternatively and the shaft
will appear as rotating with the half of actual speed of the shaft. When nf s < f sb < 0.5(2 n + 1) f s the
shaft will appear to rotate at frequency f sb nf s in the same direction as the actual. When

961
0.5(2n + 1) f s < f sb < 2 nf s the shaft will appear to rotate at frequency f sb 2 nf s in the opposite

direction as the actual. Even when we flashing frequency is higher than the Nyquist frequency the
actual shaft speed may not be seen by the stroboscope flashing because of the cyclic nature of the
shaft rotation. At f sb = 0.5nf s the shaft will appear to rotate with the half of actual speed and at
flashing frequency it will be lower than this (but the direction of rotation would change). Hence only
at the f sb = 0 we will observed the actual motion of the shaft.

16.3 Fourier Series


In mathematics, a Fourier series decomposes a periodic function into a sum of simple oscillating
functions, namely sines and cosines. The study of Fourier series is a branch of Fourier analysis.
Fourier series were introduced by Joseph Fourier (17681830) for the purpose of solving the heat
equation in a metal plate (Fourier, 1822). In data processing, we must first know the frequency
components contained in a signal. The fundamental knowledge necessary for it is the Fourier series.
We will briefly summaries it from the point of view of signal processing. On type of Fourier series is
expressed by real numbers, while the other is by complex number.
(i) Real Fourier Series:
A periodic function x(t ) with period T can be expanded by trigonometric functions which belong to
the orthogonal function systems as follows

x (t ) =

a0
+ ( an cos nt + bn sin nt )
2 n =1

(16.1)

where = 2 / T . This series is called the Fourier series or real Fourier series. Its coefficients are
given by
an =

T /2

2
x(t ) cos ntdt ,
T T / 2

bn =

T /2

2
x(t )sin ntdt
T T / 2

(16.2)

The mathematical conditions for the convergence of equation (16.1) are extremely general and cover
practically every conceivable engineering situation (Churchill, 1941). The only important restriction is
that, when x(t ) is discontinuous, the series gives the average value of x(t ) at the discontinuity.
(ii) Complex Fourier Series:

962
Fourier series can be expressed by complex numbers using Eulers formulas: e j = cos + j sin
and e j = cos j sin this gives cos =

e j + e-j
e j e-j
and sin =
. Complex numbers make it
2
2j

easier to treat the expressions. The complex representation also makes it possible to represent a
whirling motion of a rotor on the complex plane. Substituting the Eulers formula into equation (16.1)
, we have

x (t ) =
=

a0
+
2
a0
+
2

an

n =1

n =1

a
e jnt + e-jnt
e jnt e-jnt
+ bn
= 0+
2
2j
2

(c e ) + (c
n

jn t

n =1

e-jnt ) =

n=0

n =1

{0.5 ( a

jbn ) e jnt + 0.5 ( an +jbn ) e-jnt }

(c e ) + (c e ) +
n

jn t

n =1

jn t

1
n =

(c e )
n

jn t

which finally gives

c n e jnt

x(t ) =

(16.3)

n =

with

cn =

a n jbn
,
2

c0 =

a0
,
2

cn =

a n + jbn
2

(16.4)

Between the real and complex Fourier coefficients, the relationships. where the complex coefficients
are given by

1
cn =
T

T /2

x(t )e jnt dt

(n = 0,1,2, )

(16.5)

T / 2

Equation (16.3) is called the complex Fourier series. From equation (16.4), we know the following
relationship

cn = cn

(16.6)

which tells that when the real part of complex Fourier coefficients is plotted with respect to the

n(n = 0,1,2,

) , it is symmetric about the n = 0. Similarly, when the imaginary part of complex


Fourier coefficients is plotted with respect to the n(n = 0,1,2, ) , it is skew-symmetric about the
n = 0. These complex Fourier coefficients can also be represented by

963

cn = cn e j n

where the absolute value

(16.7)

c n = a n2 + bn2 2 is called an amplitude spectrum, the angle

n = c n = tan 1 (bn a n ) a phase spectrum and c n

a power spectrum.

Example 16.3 Consider a wave defined as

x (t ) = 1 for 0 t 1 and 7 t 8

= 0 for 1 t 7
The time period of the wave is T = 8. Obtain the complex Fourier coefficients of the square wave and
plot them with respect to gradually increasing Fourier series order (or harmonics) n. It should
illustrate how by considering gradual increase in harmonics of the Fourier series, it actually converges
to the real signal.
Solution: The given square wave is plotted in Figure 16.18.

Fig. 16.18 The time history of a square signal


For this square wave, we can obtain complex Fourier coefficients from equation (16.5), as follows

c0 =
and

1
T

T /2
T / 2

x(t )dt =

1
T

1
4

0dt + 1dt + 1dt +

4
1

0dt =

1
2 2
[0 + (1 + 1) + (1 0) + 0] = = = 0.25
T
T 8
(a)

964

cn =

1
T

1
4

0dt +

0
1

e j nt dt + e j nt dt +
0

4
1

0dt =

1 e j nt
T j nt

+
1

e j nt
j nt

1
(1 e jn ) + ( e jnt 1) = j1nT [ cos n jsin n + cos n jsin n ]
j nT
2sin n
=
+ j 0
Tn
(b)
for n 0
=

Complex Fourier series coefficients (i.e., the line spectrum) given by equations (a) and (b) are plotted
in Figure 16.19 for real part with respect to n. The abscissa can be taken as frequency (for n = 1 ,
f = 2 / T = 0.784 Hz; similarly for n = k , f = 2 k / T = 0.784k Hz;). This diagram is called the
spectrum of the time signal. Since given square wave is even function x(t ) = x(t ) , imaginary part of
cn is zero. For odd function we will have condition as x(t ) = x(t ) . These are illustrated in Figure

16.20.

0.2

0.15

0.1

0.05

- 0.05

- 15

-10

-5

10

15

Figure 16.19 (a) The line diagram with actual data of a square wave (complex form)

Figure 16.19 (b) The line spectrum and an envelope (dashed line) of a square wave (complex form)

965

(a)

(b)

Figure 16.20(a) An odd function (b) An even function


Fig. 16.21 shows a square wave and its various harmonic sine waves, by adding more and more these
harmonic waves the square wave could be obtained. Figures 16.22(a)-(g) show how by gradually
adding more terms in the Fourier series, it approaches the actual signal.

1.2
1
0.8
0.6
0.4

x (t )

0.2
0
- 0.2
- 0.4
- 0.6
-10

-5

Time, s
Fig. 16.21 A square wave and its various harmonics

10

966

n=1

x ( t ) 0.5
0
-9

-7

-5

-3

-1

(a)

Time, s

n=2

x ( t ) 0.5
0
-9

-7

-5

-3

-1

(b)

Time, s

n=3

x ( t ) 0.5
0
-9

-7

-5

-3

-1

(c)

7
Time, s

n=5

x ( t ) 0.5
0
-9

-7

-5

-3

-1

(d)

Time, s

967

n = 10

x ( t ) 0.5
0
-9

-7

-5

-3

-1

Time, s

(e)

n = 50

x ( t ) 0.5
0
-9

-7

-5

-3

-1

(f)

Time, s

n = 100

x ( t ) 0.5
0
-9

-7

-5

-3

-1

(g)

Time, s

Fig. 16.22 Comparison between the continuous square wave in time domain and the corresponding
complex Fourier series up to different harmonics (n is the number of harmonics included)

16.4 Fourier Transform and Fourier Integral


In previous case we have seen the spacing between adjacent harmonics is = 2 / T and it will be
seen that, when the period T becomes large, the frequency spacing becomes smaller, and the
Fourier components become correspondingly tightly packed in Figure 16.19. In the limit when
T , they will in fact actually merge together. Since in this case x(t ) no longer represents a

periodic phenomenon we can then no longer analyse it into discrete frequency components. For

968
example, when x(t ) is an isolated pulse, it cannot be converted to a discrete spectrum since it is not
periodic. Subject to certain conditions, we can however still follow the same line of thought except
that the Fourier series (16.1) (or (16.3)) turns into a Fourier integral and the Fourier coefficients (16.2)
(or (16.5)) turn into continuous functions of frequency called Fourier Transforms. Let us consider that
this interval is extended to infinity. Then the spectra obtained will represent the spectra of the isolated
pulse. Substituting equation (16.5) into equation (16.3), we get

x(t ) =

n =

1
2

T /2

x(t )e jn0 t dt e jn0t

T / 2

2
T

(16.8)

where the frequency = 2 / T of the fundamental wave is denoted by 0 and is called


fundamental harmonics. Here we represent of the nth order by n 0 = n and the difference in
frequencies between the adjacent components by n +1 n = 0 = 2 / T = . If we make

T , we have

x(t ) =

where n , and

1
2

x(t )e jt dt e jt d

(16.9)

are replaced by , d and , respectively. This can be expressed in separate

forms as follows

x(t ) =

X ( )e jt dt

(16.10)

with

1
X ( ) =
2

x(t )e jt dt

(16.11)

Equation (16.11) is called the Fourier transform of x(t ) and equation (16.10) is called the inverse
Fourier transform (or the Fourier integral) of X ( ) . The classical Fourier analysis theory (Churchill,
1941)considers the conditions that x(t ) must satisfy equations (16.10) and (16.11) to be true. For
engineering applications, the important condition is usually expressed in the form

x(t ) dt <

(16.12)

969
It means that classical theory applies only to functions which decay to zero when t . This
condition may be relaxed when impulse functions are introduced in the generalized theory of Fourier
analysis. As for the discrete Fourier series, when there is a discontinuity in x(t ) , equation (16.10)
gives average value of x(t ) at the discontinuity.

Example 16.4 Consider a square pulse defined as


x(t)

=1

for 1 t 1

=0

for all other t

Obtain the continuous spectrum for the pulse and compare the same with the square wave of example
16.3.
Solution: Figure 16.23(a) shows the square pulse given in the problem. From equation (16.11), i.e., by
the Fourier transformation, we get

1
X ( ) =
2

1
1

j t

1 e j t
dt =
2 -j

=
1

1 e j t e j t
1 2 jsin sin
=
=
2
-j
2 -j

(a)

Now, let us compare the line spectrum of a square wave of period T as shown in Figure 16.19 and that
of a square pulse shown in Figure 16.23(b) in the form of a continuous spectrum. From equation (b)
of example 16.3, we have the Fourier coefficients, cn as

cn =

2sin n0
2sin n
=
Tn =0
Tn0

Tcn sin ( n0 )
=
( n0 )
2

(b)

On comparing equation (b) with equation (a) for = 0 , the Fourier coefficients, cn , and the Fourier
transform, X ( ) , can be related as

Tc n
= X (n 0 )
2

(c)

where 0 is the fundamental frequency. Therefore, the envelope of the quantities obtained by
multiplying T / 2 to the line spectra of the Fourier coefficients, cn , of the square wave gives the

970
continuous spectra of the Fourier transform X ( ) of the square pulse. For example at = 0 from
equation (a), we have

X ( = 0) =

sin

=0

d
( sin )
cos
= d
=
d

( )
d
=0

=
=0

(d)

From equation (a) of example 16.3, we have c0 = 0.25 . Hence,

Tc0 8 0.25 1
=
= , which is same as
2
2

equation (d). Similarly, at any other harmonics it can be verified that the Fourier coefficients, cn , and
the Fourier transform X ( ) are related as in equation (c).

(a) Square pulse in time domain

(b) Fourier transform of the square pulse

Figure 16.23 A square pulse and its continuous spectrum

16.5 Discrete Fourier Transform


Till now, we have discussed the Fourier series, Fourier transform and Fourier Integral on the
assumption that we know a continuous signal wave (it is also called the continuous time series) in the
infinite time domain. However, in practical experiments, the data acquired, converted from the data
measured by an analog-to-digital converter, are sequences of data {xn } (n = 0, 1, 2, , n-1, n) that are
discrete and with finite number. To perform the spectrum analysis using these finite numbers of
discrete data (it is also called discrete time series), we must use the discrete Fourier transform (DFT).
In order to estimate spectra from discrete Fourier series the obvious method is to estimate the
appropriate correlation function (e.g., the auto- and cross- correlation functions) first and then to
Fourier transform this function to obtain the required spectrum. Until the late 1960s, this was the basis
of practical calculation procedures which followed the formal mathematical route by which spectra
are defined as Fourier transforms of correlation functions. The assumptions and approximations

971
involved were studied in detail and there is an extensive literature on the classical method (Bendat, et
al. 1966). However the position was changed by the advent of the fast Fourier transform (or FFT).
This is a remarkably efficient way of calculating the Fourier transform of a discrete time series.
Instead of estimating spectra by first determining correlation of a time series and then calculating their
Fourier transforms, it is now quicker and more accurate to calculate spectral estimates directly from
discrete time series by a method which we shall describe in detail.
This DFT is defined as follows: Given N data sampled with a constant interval t , the DFT is
defined as a series expansion on the assumption that the original signal is a periodic function with the
period Nt (although the original signal is not necessary periodic). However, various problems occur
in the course of this processing. On performing FT in a discrete environment introduces artificial
effects, like aliasing effects, spectral leakages, scalloping losses, etc.
If the sampling rate, in the time domain, is lower than the Nyquist rate the aliasing occurs. Two
signals are said to alias if the difference of their frequencies falls in the frequency range of interest,
which is always generated in the process of sampling (aliasing is not always bad; it is called mixing or
heterodyning in analog electronics, and is commonly used in tuning radios and TV channels). It
should be noted that, although obeying the Nyquist sampling criterion is sufficient to avoid aliasing, it
does not give high quality display in time domain record. If a sinusoid existing in the time signal not
bin-centered (i.e., if its frequency is not equal to any of the frequency samples) in the frequency
domain spectral leakage occurs. In addition, there is a reduction in coherent gain if the frequency of
the sinusoid differs in value from the frequency samples, which is termed scalloping loss.
(a) The first is the aliasing problem. When the signal is sampled with a interval t , the information
about the components with frequencies higher than

1
is lost, which is the Nyquist frequency.
2 t

Therefore, we must only consider to the valid range of the spectra obtained, i.e., below the Nyquist
frequency.
(b) The second is the challenge of the coincidence of periods. It is impossible to know the correct
period of the original signal before the measurement. Therefore, the period of the original signal and
the period of DFT do not coincide, and this difference produces the leakage error. We will discuss this
leakage error and its countermeasure subsequently (e.g., by window functions).
(c) The third is the problem about the length of measurement. In the case of an isolated signal x(t), we
cannot have data in an infinite time range. However, since the Fourier coefficients c n and Fourier

972
transform X ( ) coincide at discrete points as explained in previous section, we can obtain X ( ) by
connecting the values of c n smoothly.
Now the procedure of computing the DFT is explained in the subsequent section. Let us assume that
we obtained discrete time series x 0 , x1 , x 2 ,

, x N 1 by the sampling. These data are extended

periodically to make a virtual periodic signal, as shown by the dashed curve in Figure 16.24.

Figure 16.24 Formation of virtual continuous period signal with sampled signal sequences
The fundamental period is T = Nt and the fundamental frequency is 0 = = 2 T . If this
dashed curve is given as a continuous time function, its Fourier series expansion is given by the
expressions obtained by replacing with in equation (16.3) and (16.5). However, in the case of
a discrete signal, the integral of equation (16.5) must be calculated by replacing t, T, x(t) and with

k t , N t , xk and

respectively. By such replacements, we have

cn =

1
T

N 1
k =0

xk e jn k t t =

1
N

N 1
k =0

xk e jn (2 / T ) k (T / N )

(16.13)

We represent the right-hand side of this expression by X n , that is

Xn =

1
N

N 1
k =0

xk e j2 nk / N ; (n = 0, 1, 2, , N-1)

and call this descrete Fourier transform of the discrete time signal x 0 , x1 ,

(16.14)

, x N 1 . Paired with this is

the following expression, called the inverse discrete Fourier transform (IDFT)

973
xk =

N 1
n=0

X n e j2 nk / N ;

(k = 0, 1, 2, , (N-1))

(16.15)

These transformations map the discrete signal of a finite number on the time axis to the discrete
spectra of a finite number on the frequency axis, or vice versa. These expressions using complex
numbers are called the complex discrete Fourier transform and the complex inverse discrete Fourier
transform. We also have transformation using only real numbers. One is the real discrete Fourier
transform, given by

An =

1
N

N 1

1
N

N 1

k =0

xk cos

2 nk
N
(n = 0, 1, 2, , N-1)

Bn =

k =0

xk sin

(16.16)

2 nk
N

where An and Bn are quantities defined by X n = An + jBn . Further, the inverse real discrete
Fourier transform is given by

xk =

N 1
n=0

An cos

2 nk
2 nk
Bn sin
;
N
N

(n = 0, 1, 2, , N-1)

(16.17)

We explain the characteristics of the spectra obtain by the DFT using an example subsequently.
Before that let us examine the aliasing effect on the DFT. We have seen that the DFT of the series

x 0 , x1 ,

, x N 1 is defined by

Xn =

1
N

N 1
k =0

xk e j2 nk / N ;

(n = 0, 1, 2, , N-1)

(16.18)

Suppose we try to calculate values of X n for the case when n is greater than (N-1). Let for example,
n = N + 1 . Then,

X N +1 =

1
N

N 1
k =0

xk e j(2 k / N )( N +1) =

1
N

N 1
k =0

xk e j(2 k ) e j(2 k / N ) =

1
N

N 1
k =0

xk e j( 2 k / N ) = X 1

Since e j(2 k ) is always equal to 1 whatever the value of k. Hence, we have

(16.19)

974
X N +1 = X 1 ,

X N +2 = X 2 ,

(16.20)

The coefficients X n therefore just repeat themselves for n > ( N 1) , so that if we plot the magnitudes
X n along a frequency axis n = 2 n / N t , the graph repeats itself periodically. Furthermore, it is
also easy to see that from equation (16.18) that, provided terms in the { xk } series are real, we have

X 1 =

1
N

N 1
k =0

xk e j2 k / N = X 1*

(16.21)

where X1* is the complex conjugate of X 1 . Hence


X 1 = X 1

(16.22)

so the spectra diagram for the X n would be symmetrical about the zero frequency position. The
unique part of the graph occupies the frequency range / t . Higher frequencies just show
spurious Fourier coefficients which are repetitions of those apply at frequencies below / t . We can
therefore see that the coefficients X n calculated by the DFT are only correct Fourier coefficients for
frequencies up to

n =

2 n
=
N t t

n=

that is for n in the range n = 0,1, 2,

N
2

(16.23)

, N / 2 . Moreover, if there are frequencies above / t present in

the original signal, these introduce a distortion of the graph called aliasing. The high frequency
components contribute to the { xk } series and falsely distort the Fourier coefficients calculated by the
DFT for frequencies below / t . If max is the maximum frequency component present in x(t ) ,
then aliasing can be avoided by ensuring that the sampling interval t is small enough that

> max

with

f max =

or

max
2

1
> f max
2 t

(16.24)

(16.25)

975
The frequency 1/ 2 is called the Nyquist frequency (or some times the folding frequency) and is the
maximum frequency that can be detected from data sampled at time spacing . The phenomenon of
aliasing is most important when analyzing practical data. The sampling frequency 1/ 2 must be high
enough to cover the full frequency range of the continuous time series. Otherwise the spectrum from
equally spaced sampled will differ from the true spectrum because of aliasing. In some cases the only
way to be certain that this condition is met may be to filter the time series to remove intentionally all
frequency components higher than 1/ 2 before beginning the analysis.

Example 16.5 A square wave with period T = 8 sec and sixteen sampled data: x 0 =

x5 =

= x 4 = 1 and

= x15 = 0 . (i) Obtain the spectrum of the discrete time series. (ii) Compare the spectrum of

the same square wave for thirty two sampled data: x0 =

= x8 = 1 and x9 =

= x31 = 0 . It should be

noted that the signal is sampled intentionally in the range that coincides with the period of the original
square wave to avoid the leakage error.
Solution: (i) Sixteen sampled points: The sampling of the square wave for sixteen sampled data is
shown in Figure 16.25. The period of the signal is 8 sec, hence the fundamental frequency would be
f 0 = 1/ T = 1/ 8 = 0.125 Hz. The sampling interval is t = T / N = 8 /16 = 0.5 sec. The Nyquist

frequency would be f c = 1/ 2t = 1/ ( 2 0.5 ) = 1 Hz. Hence, the maximum harmonics which will be
valid is n = f c / f 0 = 1/ 0.125 = 8 . Total number of harmonics are same as number of sampled points,
i.e., 16.

Figure 16.25 Discrete time series of a square wave


On using equation (16.14), the spectrum, X n , can be calculated by simple substitution and addition of
various terms. For the present case N = 16 . In general, X n is a complex quantity and it can be plotted
in various form, e.g., (1) the real An = Re ( X n ) and imaginary Bn = Im ( X n ) parts versus harmonic
number, n, (or frequency, f), (Figure 16.26(a, b)) and (2) the magnitude X n and phase X n versus
harmonic number, n, (or frequency, f) (Figure 16.26(c, d)).

976

(a) Real part of the spectrum

(c) Magnitude part of the spectrum

(b) Imaginary part of the spectrum

(d) Phase part of the spectrum

Figure 16.26 Spectrum of the discrete time series for a square wave with sixteen sampled points

Figure 16.27 Spectrum of the discrete time series for a square wave with thirty-two sampled points
These spectra have the following characteristics: (a) The spectra is periodic with period N (this is also
observed in equation (16.20)). (b) The same spectra as those of the negative order n = -N/2, , -1 also
appear in the range n = N/2, , (N-1). (c) The spectra of the real part and those of the amplitude are
both symmetric about n = N/2. (d) The spectra of the imaginary part and those of the phase are skewsymmetric about n = N/2. (e) The spectra in the left half of the zone n = 0, , (N-1) are valid. The
spectra in the right half are virtual and are too high compared to the sampling frequency (this is also
observed while discussing the aliasing effect in equation (16.23)).

977

(ii) Thirty-two sampled points: The period of the signal is same as 8 sec, hence the fundamental
frequency would also be same as f 0 = 1/ T = 1/ 8 = 0.125 Hz. The sampling interval is now
t = T / N = 8 / 32 = 0.25 sec. The Nyquist frequency would be f c = 1/ 2t = 1/ ( 2 0.25 ) = 2 Hz.

Hence, the maximum harmonics which will be valid is n = f c / f 0 = 2 / 0.125 = 16 . Total number of
harmonics are same as number of sampled points, i.e., 32.
When the sampling interval is narrowed (from 0.25 sec to 0.125 sec) the number of spectra increases
(from 16 to 32) as shown in Figure 16.27, and therefore such a spectra diagram written in the interval

= 2 / T (which remains constant since T is constant) extends to the right (from 2 Hz to 4 Hz).
An envelope is shown by dashed line of the spectra in Figure 16.27. If the sampling frequency is
shortened continuously, the sampled data become substantially equal to the continuous wave, and
therefore its spectra will approach those of the Fourier series shown in right half of Figure 16.23(b).
The magnitude of X 0 is 0.313 in Figure 16.26(a) and X 0 is 0.281 in Figure 16.27. This value
approaches c0 = 0.25 in Figure 16.23(b) as the number of data sampled increases.
Different types of definition of DFT and IDFT are used, depending upon personal preference. Some
use the following definitions, in which the magnitudes of X n coincide with that of Fourier transform
X ( ) in Figure 16.23(b).

X n = t

N 1
k =0

xk e j2 nk / N

(n = 0, 1, 2, N-1)

(16.26)

(k = 0, 1, 2, N-1)

(16.27)

and

xn =

1
T

N 1
n =0

X n e j2 nk / N

Some use following expressions, which have the coefficient 1/N in the counter-part expression:
(MATLAB uses this)

Xn =

N 1
k =0

xk e j2 nk / N

(n = 0, 1, 2, N-1)

(16.28)

(n = 0, 1, 2, N-1)

(16.29)

and

xn =

1
N

N 1
n =0

X n e j2 nk / N

978
Of course, every definition has the same function as mapping from time domain to frequency domain.
Especially, when we are interested in the critical frequencies at which amplitudes have peaks, then all
the definitions can be used equally well. However, we must be careful when we interpret the physical
meaning of the magnitude of the spectra. For example, for x (t ) = sin t , it is equation (16.14) that
gives a spectrum with magnitude 1.

16.6 Fast Fourier Transform


The vast computational task necessary for DFT prevented its practical utilization. In 1965, Cooley and
Tukey proposed a computer algorithm that enabled the fast computation of DFT. The algorithm is
called Fast Fourier Transform (FFT), has made real-time spectrum analysis a practical tool. In the
calculation of the DFT given by equation (16.18), as

Xn =

1
N

N 1
k =0

xk e j2 nk / N ;

(n = 0, 1, 2, , N-1)

If we were working out values of X n by a direct approach we should have to make N multiplications
of the form ( xk ) ( e j2 nk / N ) for each of N values of X n and so the total work of calculating the full
sequence X n would require N 2 multiplications. However, the same calculation appears repeatedly
since the function e j2 nk / N = cos {nk ( 2 N )} jsin {nk ( 2 N )} has a periodic characteristic. The
FFT algorithm eliminated such repetition and allowed the DFT to be computed with significantly
fewer multiplications than direct evaluation of DFT. The FFT reduces this work to a number of
operation of the order of 2 N log 2 ( N ). The FFT algorithm has the restriction that the number of data
must be 2

( n = 1,

2,

, ) . This allows certain symmetries to occur reducing the number of

calculations (especially multiplications) which have to be done. When the number of data N is

N = 2 n , DFT needs N 2 multiplications and FFT needs 2 Nn multiplications, which is only 2 n / N


of the number of operations. For example, when = 211 = 2048 , about 4,194,304 multiplications are
necessary in DFT and about 45,056 in FFT, which is only about 1/93rd of number of operations. If N
increases this difference increases extremely large. The FFT therefore offers the added bonus of an
increase in accuracy. Since fewer operations have to be performed by the computer, round-off errors
due to the truncation of products are reduced, and the accuracy increases. If the length of the discrete
time series is not equal to 2

( n = 1,

2,

, ) , then the discrete time series are zero padded (add

zeros in the series as discrete data) till it reaches up to the next 2n value. For example, if number of
discrete data are 1000, then we need to zero-pad 24 more data to make it to N = 1024 = 210 . This will
not affect the accuracy of the FFT at all. For example, if we have 520 discrete data then instead of
zero-padding 1024-520 = 504 data, the original data may be truncated to next lower value of

979
29 = 512 . This may introduce some error due to truncation of the original data, howver, if the discrete
time series is long enough then error introduced will be acceptably small. However, the state of the art
analog-to-digital equipments now digitizes data 2

( n = 1,

, ) in number. MATLAB has FFT

2,

function name fft ( x n ) where x n = {x}n1 .


Example of zero padding/truncation to be included.

16.7 Leakage Error and Countermeasures

Measured
Range case A

Measured
Range case B

1.0

-1.0

10

12

(a) Original signal and measured range

6 2

6
4

10

2
0

2 4

(b) Assumed signal and its spectrum (case A)

4 6

8 10 k 12

16

8 10

4
2
0

6 8 10 12 14 16

(c ) Assumed signal and its spectrum (case B)


Figure 16.28 Leakage error

In FFT or DFT, computations are based on the assumption that the data sampled over a time
period are repeated before and after data measurement. Figure 16.28 shows the assumed signals
and their spectra for two types of measurement of a sinusoidal signal x (t ) = sin t . Both cases
have 32 sampled data, but their sampling intervals are different. In case A, the sampling interval is
t = 4 / 32 = 0.3927 and the range measured is exactly twice the fundamental period. The

980
computation of FFT or DFT is performed for the wave as shown by the dotted line (Figure
16.28(b). In this case the assumed wave is same as the original signal and therefore we get a
correct signal spectrum. In case B, the sampling interval is t = 5 / 32 = 0.4909 , and the range
measured is about 2.5 times the period of the original signal. In this case, the assumed wave
shown in Figure 16.28(c) is not smooth at the junction and differs from the original signal in time
domain. As a result, the magnitude of the correct spectrum decreases and spectra that do not exist
in the original signal appear. As seen in this example, if the time duration measured and the
period of the original signal do not coincide, the magnitude of the correct spectrum decreases and
spectra that do not exist in the original signal appear on both sides of the correct spectrum. This
phenomenon is called the leakage error.

Example 16.6 Illustrate the leakage error with the help of a simple sine wave. Take two fullcycles as sampled signal and then take various sampled signals between two full-cycle and twoand-half cycles.
Answer: Figure 16.29(a) show a continuous sine wave and its FFT. Figs. 16.29(b-f) shows
sampled signal of the continuous sine wave at various time length and corresponding FFT. It can
be seen that except 16.29(b) all others have leakage error. The leakage error is seen to increase
with the discontinuity at the junction point of two sampled signal for FFT (see 16.29(c-f)).

2000

FFT [x(t)]

x(t)

0.5
0

-0.5
-1
0

2pi

4pi

6pi

Time, s

8pi

1500
1000
500
0

10

20

30

Frequency

(a) Original continuous sinusoidal signal and its FFT

40

50

981
1500

FFT[x(t)]

x(t)

0.5
0

-0.5
-1
0

2pi

4pi

1000
500
0

6pi

10

15

20

25

Time, s
Frequency
(b) Sampled sinusoidal signal (case 1) and its FFT
1500

FFT [x(t)]

x(t)

0.5
0

-0.5
-1
0

2pi

4pi

6pi

1000

500

8pi

10

15

20

25

15

20

25

Frequency

Time, s
(c) Sampled sinusoidal signal (case 2) and its FFT
1500

FFT [x(t)]

x(t)

0.5
0

-0.5
-1
0

2pi

4pi

6pi

1000
500

8pi

10

Time, s
Frequency
(d) Sampled sinusoidal signal (case 3) and its FFT
1500

FFT [x(t)]

x(t)

0.5
0

- 0.5
-1
0

2pi

4pi

6pi

8pi

1000

500

10

15

Time, s
Frequency
(e) Sampled sinusoidal signal (case 4) and its FFT

20

25

982
1500

FFT [x(t)]

x(t)

0.5
0

-0.5
-1
0

2pi

4pi

6pi

1000

500

8pi

10

15

20

25

20

25

Frequency
(f) Sampled sinusoidal signal (case 5) and its FFT

Time, s

1500

1000

FFT [x(t)]

x(t)

0.5
0
- 0.5
-1
0

2pi

4pi

6pi

8pi

500

Time, s

10

15

Frequency

(g) Sampled sinusoidal signal (case 6) and its FFT


Figure 16.29 A continuous sine wave and its sampled signals with corresponding FFTs
(ii) Countermeasures for leakage error (Window Function):
There is a difficulty if the time signal is not periodic in the time record, especially at the edges of the
record (i.e., window). If the DFT or FFT could be made to ignore the ends and concentrate on the
middle of the time record, it is expected to get much closer to the correct signal spectrum in the
frequency domain. This may be achieved by a multiplication by a function that is zero at the ends of
the time record and large in the middle. This is known as windowing.
It should be realized that, the time record is tempered and perfect results shouldn'
t be expected. For
example, windowing reduces spectral leakage but does not totally eliminate it. It should also be noted
that, windowing is introduced to force the time record to be zero at the ends; therefore transient
signals which occur (starts and ends) inside this window do not require a window. They are called
self-windowed signals, and examples are impulses, shock responses, noise bursts, sine bursts, etc.
To decrease the leakage error due to discrepancy between the time duration measured and the period
of the original signal, we must connect the repeated wave smoothly. For this purpose we multiply a
weighting function that decrease gradually at both sides. This weighting function is called time

983
window. Representative time windows: the Hanning window, Hamming window and BlackmanHarris window are shown in Figure 16.30. These windows are defined in the range: 0 n N 1 as
Rectangular window:

w( n) = 1.0

Hanning window:

w(n) = 0.5 0.5cos ( 2 n / N )

Hamming window

w(n) = 0.54 0.46 cos ( 2 n / N )

Gaussian:

w ( n ) = exp 0.5 ( 2 n / N )

Blackman-Harris window:

w(n) = 0.423 0.498cos ( 2 n / N ) + 0.0792 cos ( 4 n / N )

and outside 0 n < N 1

( n) = 0 for all cases.

(16.30)

Various window functions with a typical signal with discontinuity at the junction point are shown in
Figs. 16.31-16.33. Effects on the leakage error by window functions on the spectrum are shown in Fig
16.34 (a-d).

Hanning window
Hamming window
Blackman-Harris
window

Weighing function, w

0.8

0.6

0.4

0.2

200

400

600

800

Number of data points, N


Figure 16.30 Various representative window functions

1000

984

x(t)

0.5

-0.5

w(n)

-1
0

10

15

20

25

30

20

25

30

20

25

30

Time, s
Fig 16.31 With the Hanning window

x(t)

0.5

- 0.5

w(n)

-1
0

10

15

Time, s
Fig. 16.32 With the Hamming window

x(t)

0.5

-0.5

w(n)

-1
0

10

15

Time, s
Fig. 16.33 With the Blackman-Harris window

985
1500

FFT[x(t)]

x(t)

0.5
0

-0.5
-1
0

2pi

4pi

6pi

10

20

25

20

25

20

25

1500

FFT[x(t)]

1000

500

-1
0

2pi

4pi

6pi

8pi

10pi

Time, s

10

(b) The signal with the Hanning window

15

Frequency

1500

FFT[x(t)]

x(t)

15

Frequency

(a) Original signal

x(t)

500

8pi

Time, s

1000

-1
0

2pi

4pi

6pi

Time, s

8pi

1000
500
0

10pi

10

15

Frequency

(c) The signal with the Hamming window


1500

FFT[x(t)]

x(t)

-1
0

2pi

4pi

6pi

Time, s

8pi

10pi

1000
500
0

10

15

Frequency

20

25

(d) The signal with the Blackman-Harris window

Fig. 16.34 Illustration of leakage error and application of various window functions on a same signal

986
For a discussion of their characteristics and the effects of these window functions, refer to some books
on signal processing (Bendat and Piersol, 2010).
(iii) Prevention of leakage by coinciding periods: As discussed above, we can obtain the correct result
if the time duration measured coincides with the integer multiple of the period of the original signal. If
we can attain this by some means, it is better than the use of window functions, which distorts the
original signal. For example, for numerical calculations that can be repeated in exactly the same way
and whose sampling interval can be adjusted freely, we can determine the measurement duration after
we know the period of the original signal by trial simulation, and then execute the actual numerical
simulation. On the contrary, for experiments, fine adjustment of sampling intervals is generally
impossible using practical measuring instruments. However, if the phenomenon appears within a
speed range, we can change the rotational speed little by little and adopt the best result where the
period, often determined by the rotational speed, and the sampling interval fit.

16.8 Applications of FFT to Rotor Vibrations


In the investigation of rotor vibration, we must know the direction of a whirling motion as well as its
angular velocity. In FFT (or DFT), elements of data sequence

{x k }

obtained by sampling are

considered as real numbers and those of data sequence {X n } obtained by discrete Fourier transform
are considered as complex numbers. In the following, we introduce a method that can distinguish
between whirling directions utilizing the revised FFT. In this FFT, rotor whirling motion is
represented by a complex number by overlapping the whirling plane on the complex plane and
applying FFT to these complex sampled data. Let us assume that a disc mounted on a elastic shaft is
whirling in the y-z-plane. We get sampled data {y k } and

{z k } by measuring the deflections

y (t )

and z (t ) in the y and z directions respectively. Taking the y-axis as real axis and the z-axis as
imaginary axis, we overlap the whirling plane on the complex plane. Using sampled data y k and z k ,
we define the complex numbers as follows:

S k = y k + jz k

(k = 0, 1, 2,, N-1)

and apply FFT (DFT) to them. We call such a method the complex-FFT method.

(16.31)

987
Fig. 16.35(a) shows sub-harmonic vibration component of order in vibration signal. In
Figs.16.35(b) and (c), the same spectrum are shown with windowing and tuning of sampling interval,
respectively, for clarity of the frequency components. Fig. 16.36 shows the complex-FFT with distinct
forward and backward whirl frequency components.

Subharmonic resonance + / 2
+
of order of a forward

+ / 2
-2
-1
Backward

-2

2
Forward

(a) Spectrum without window

(b ) Spectrum with window

+ / 2

-2

-1

-1

(c) Spectrum obtained by adjusting sampling interval


Figure 16.35 Spectra of the sub-harmonic resonance of order of a forward whirling mode

-2
Backward

-1

Forward

Figure 16.36 Spectrum of the combination resonance (complex FFT method)

988
16.9 Properties of Random Discrete Signals
16.9.1 Probability, Probability Distribution Function, and Probability Density Function
A probability can be defined as simply the fraction of favourable events out of all possible events.
Probabilities are inherently non-negative; i.e., they can be only be positive or zero. Neither
assumption is strictly true in practice but both provide useful engineering solutions when they are
suitably interpreted. The measure of probability used is based on a scale such that the probability of
the occurrence of an event which cannot possibly occur is taken to be zero; the probability of
occurrence of an event which is absolutely certain to occur is taken to be unity. Any other event
clearly must have a probability between zero and unity. Suppose that we spin a coin: the probability
of the result heads H is equal to that of the result tails T, which will be equal to
Pr[ H ] = Pr[T ]

(16.32)

where Pr[ ] represents the probability. As the probability of either heads or tails, which will be equal
to Pr[ H ] + Pr[T ] , must be unity, i.e.,
Pr[ H ] + Pr[T ] = 1

(16.33)

From equations (16.32) and (16.33), we get


Pr[ H ] = Pr[T ] = 12

(16.34)

Similarly, the probability of throwing any given number with a symmetrical six-sided die would be
1/ 6 , since all numbers from 1 to 6 are equally probable and their total probability must be unity. The

probability of throwing an odd number with the die is

1
2

, since Pr[Odd] + Pr[Even] = 1 and

Pr[Odd] = Pr[Even] .
Now define a probability of throwing a given number by the die as
Pr[ N = n] = p (n) = 1/ 6 ,

1 n 6

(16.35)

The probability of throwing the die that N is odd

Pr[ N = 1 or 3 or 5] = p(1) + p(3) + p(5) =

1 1 1 1
+ + =
6 6 6 2

(16.36)

989

which we obtained earlier also. Hence, it can be generalized as

Pr[ N n] =

p (r ) = P (n)

(16.37)

The quantities p(n) and P(n) provide alternative means of describing the distribution of probability
between the various possible values of n. Definition of p(n) and P(n) are given in equations (16.35)
and (16.37), respectively.
The expectation, E ( N ) , is the expected result in any given trial, assumed to be equal to the mean
result of a very large number of trials. In the case of the die we can assume that in many throws all the
numbers from one to six will recur with equal frequency, so the expectation here is the mean of 1, 2,
3, 4, 5, 6, i.e.,
E ( N ) = (1 + 2 + 3 + 4 + 5 + 6) / 6 = 3.5

(16.38)

The above expectation can also be obtained as weighted mean of the numbers 1 to 6, as (with
p(n) = 1/ 6 and n = 6 ), i.e.,

E(N ) =

np (n)

(16.39)

The above discussions of the coil and the die involves discrete values, however, the results obtained
can be extended for a continuous values also. Suppose we have a large number of displacement data
of different amplitudes. For this example, Pr[ X = x] = 0 , i.e., the probability that a chosen
displacement value X will be equal to a fixed displacement x will be zero. However, it will be
appropriate to use P ( x) = Pr[ X x] , i.e., the probability that the chosen displacement x is below a
certain displacement value X. This quantity , P( x) , is known as the probability distribution function.
Figure 16.37 shows a typical displacement signal and Figure 16.38 shows the corresponding
probability distribution function.

990

Figure 16.37 A typical random signal

Figure 16.38 The probability distribution function

Figure 16.39 The probability density

991
As the probability can be added, they can also be subtracted. The probability of any randomly selected
displacement having a magnitude between two limits x1 and x2 can be written as
Pr[ x1 X x2 ] = Pr[ X x2 ] Pr[ X x1 ] = P ( x2 ) P ( x1 )

(16.40)

The probability that X lies between x and x + dx can be written as


Pr[ x X x + dx] = Pr[ X x + dx] Pr[ X x] = P ( x + dx) P ( x) = dP( x)

(16.41)

For dx to be infinitesimal, we have

dP ( x) = p ( x)dx

or

p ( x) =

dP( x)
dx

(16.42)

where p( x) is called the probability density function, which is an alternate way of describing the
probability distribution of a random variable, x. Figure 16.39 shows the probability density of the
random displacement signal shown in Figure 16.37. Equation (16.42) can be expressed as

P( x) =

p ( s )ds

(16.43)

Suppose we want to know the steady and fluctuating component of our signal. The steady component
is simply the mean value or expectation E[ x(t )] of x(t ) . Noting equation (16.39), on similar analogy
we can write

E[ x(t )] =

xp ( x)dx

(16.44)

where square brackets indicate the ensemble average of the quantity. It can be seen that E[ x(t )] is
given by the position of the centroid of the p( x) diagram, since

p ( x)dx = 1

(16.45)

992
16.9.2 Random Process, Ensemble, and Sample Function
The central notion involved in the concept of a random process x(t ) is that not just one time history
is described but the whole family or ensemble of possible time histories which might have been the
outcome of the same experiment are described. Any single individual time history belonging to the
ensemble is called a sample function. A random process can have several sample functions x( j ) (t ) (
j = 1, 2,

, n ) defined in the same time interval (see Fig. 16.40). Each x( j ) (t ) is a sample function of

the ensemble. Hence, a random process can be thought of as an infinite ensemble of sample functions.

Figure 16.40 Ensemble of sample functions x( j ) (t )

16.9.3 Stationary and Ergodic Process


Often the random process is simplified and is assumed to be stationary process. This assumption is
analogous to the assumption of steady state forced response in deterministic vibration.
Displacement, x(t ) , at a particular location in test rig measured under identical conditions for n
number of times will have the following form: x(1) (t ), x (2) (t ),

, x( m ) (t ) { x(t )} , which is called the

random process. These are the same physical quantities, however, they will not themselves be
identical but will have certain statistical properties in common. A random process is said to be
stationary if its probability distributions are invariant under a shift of the time scale of the signal. For
example, the random process will be stationary when displacements x(1) (tn ), x(2) (tn ),

, x( m ) (tn ) have

993
probability distributions independent of time tn ( n = 1, 2,
distribution of x(1) (tn ), x(2) (tn ),

).

By ergodic we mean the probability

, x( m ) (tn ) at any one time is equal to the probability distribution of

any one displacement signal x( j ) (t ) with respect to time. Hence, an ergodic signal will be stationary,
however, a stationary signal will not necessarily be ergodic. That means ensemble and temporal
averages of the ergodic signal will be same, whereas, these averages are different for the stationary
process. It should be noted that for the ergodic process a signal will be enough to define the statistical
properties of the process completely.
For stationary process, in particular the probability density p( x) becomes a universal distribution
independent of time. This implies that all the averages based on p( x) (e.g., the mean E[ x] and the
variance 2 = E x 2 ( E[ x ]) ) are constants independent of time. The autocorrelation function is
2

defined as
E[ x(t ) x(t + )] = R ( )

(16.46)

which is also independent of t and function of time lag = t2 t1 (for present case t2 = t + and t1 = t
). It should be noted that R (0) reduces to the mean square E x 2 . In case x has zero mean, E[ x] = 0 ,
then the mean square is identical with the variance and R(0) = 2 . The stationary assumptions can be
verified for experimental signals by calculating the mean and auto-correlation functions at different
times and checking for its invariance.
There are certain properties of autocorrelation function and its derivatives which are useful in
analysis. The autocorrelation function is an even function, since for stationary random process we can
write
R ( ) = E[ x(t ) x(t + )] = E[ x(t ) x(t )] = R ( )

(16.47)

Using a prime to indicate differentiating with respect to the contents of a bracket,


R ( ) = E[ x (t ) x(t + )] = E[ x (t ) x (t )] = R ( )

(16.48)

For = 0 , we have E[ x (t ) x (t )] = E[ x(t ) x (t )] , which is only true when both are zero. Hence,
R (0) = 0

(16.49)

994
Similarly,

R( ) =

E[ x(t ) x(t + )] =
E[ x(t ) x(t )] = E[ x(t ) x(t )]

(16.50)

For = 0 , we have

R(0) = E[ x(t ) x(t )] = E[ x2 (t )]

(16.51)

16.9.4 Probability Distribution and Probability Distribution Function


Many naturally occurring random vibrations have the Gaussian probability distribution, which is
defined as

1
p ( x) =
e
2

( x m )2
2 2

(16.52)

where m and represent mean and variance, which are constants for a particular random process.
Fig. 16.41 is a comparison of probability density with Gaussian approximation and the actual
distribution (without smoothening) from signal in Fig. 16.37. Fig. 16.42 shows the comparison of
probability density with Gaussian approximation and the actual distribution (with smoothening).

Figure 16.41 Gaussian probability density versus actual probability density without smoothening

995

Figure 16.42 Gaussian probability density versus smoothened actual probability density

16.9.5 Ensemble Average, Temporal Average, Mean, Variance


The variance is defined as the mean-square value of the difference from the mean value. Thus

2 = E

x(t ) E [ x(t ) ]

= E x 2 (t ) 2 x(t ) E [ x(t )] + E [ x(t )]

(16.53)

= E x (t ) 2 E [ x(t )] + E [ x(t ) ] = E x (t ) E [ x(t ) ]


2

where E x 2 (t )

is the square mean and E [ x(t )] is the mean square. It should be noted that as
2

E [ x(t )] is given by the first moment of area of the probability density curve about the p( x) axis, so

2 which by equation (16.53) is the expectation of x (t ) E [ x(t ) ]

is given by the second moment

of area about x = E[ x(t )] , i.e.

2 =

[ x E[ x(t )]]

p ( x)dx

(16.54)

is the radius of gyration of the probability density curve about x = E[ x(t )] . If the mean value is
zero the variance is given by

996
2 =

x 2 p ( x)dx

(16.55)

To evaluate ensemble averages, it is necessary to have information about the probability distribution
of the samples or at least a large number of individual samples. Given a single sample x ( j ) of duration
T it is, however, possible to obtain averages by averaging with respect to time along the sample. Such
an average is called a temporal average in contrast to the ensemble or statistical averages described
previously. The temporal mean of x(t )

1
< x(t ) >=
T

T /2

x(t )dt

(16.56)

T / 2

and the temporal mean square is

1
< x (t ) >=
T

T /2

x 2 (t )dt

(16.57)

T / 2

where the angular brackets represent the temporal average. When x(t ) is defined for all time
averages are evaluated by considering the limits as T . For such a function a temporal
autocorrelation ( ) can be defined

1
( ) =< x(t ) x(t + ) >= lim
T T

T /2

x(t ) x(t + )dt

(16.58)

T / 2

When ( ) is defined for a finite interval, then similar expression can be used by carefully choosing
the limits. Note that when (0) reduces to the temporal mean square. Within the subclass of
stationary random process a further subclass known as ergodic process. An ergodic process is one for
which ensemble averages are equal to the corresponding temporal averages taken along any
representative sample function. Thus for an ergodic process x(t ) with samples x( j ) (t ) we have

E[ x] =< x ( j ) >

and

R ( ) = ( )

(16.59)

An ergodic process is necessarily stationary since < x ( j ) > is a constant while E[ x] is generally a
function of time t = t1 at which the ensemble average is performed except in the case of a stationary

997
process. A random process can, however, be stationary without being ergodic. Each sample of an
ergodic process must be completely representative of the entire process.

16.9.6 Auto-correlation Function and Covariance


The autocorrelation is defined as

E[ x(t1 ) x(t2 )] =

x x p( x1 , x2 ) dx1dx2

1 2

The prefix auto refers to the fact that x1 x2 represents a product of values on the same sample at two
instants. For fixed t1 and t2 this average is simply a constant; however, in subsequent applications t1
and t2 will be permitted to vary and the autocorrelation will in general be a function of both t1 and t2 .
In an important special case the autocorrelation function is a function only of = t2 t1 .
A related average, the covariance is obtained by averaging the product of the deviation from the
means at two instants. Thus, we have the covariance as

(x

E ( x1 ) ) ( x2 E ( x2 ) ) =

(x

E ( x1 ) ) ( x2 E ( x2 ) ) p ( x1 , x2 ) dx1dx2

= E [ x1 x2 ] E [ x1 ] E [ x2 ]

(16.60)

When x1 and x2 have zero means, the covariance is identical to the auto-correlation. When t1 = t2 , the
covariance becomes identical with the mean square.

A frequency decomposition of the R ( ) can be made in the following way

R ( ) =

S ( )e jt d

(16.61)

where S ( ) is the Fourier transform of R ( ) , except for the factor 2 . A physical meaning can be
given to S ( ) by considering the limited case of equation (16.61) in which the time shift = 0 is
taken

R (0) = E x 2 =

S ( )d

(16.62)

998

The mean square of the process equals the sum over all frequencies of S ( )d so that S ( ) can be
interpreted as a mean square spectral density. It should be noted that the dimensions of S ( ) are
mean square per unit of circular frequency. Note that according to equation (16.62) both negative and
positive frequencies are considered, which is convenient for analytical investigations. In experimental
work a different unit of spectral density is widely used. The difference arises due to the use of cycles
per unit time (Hz) in place of rad/s and due to considering only positive frequencies. The experimental
spectral density will be denoted as W ( f ) where f is frequency in Hz.

R (0) = E x 2 =

S ( )d = 2

S (2 f )d ( 2 f ) = 4

W ( f )df

(16.63)

with

= 2 f

(16.64)

From above equation, the relationship between S ( ) and W ( f ) is simply


W ( f ) = 4 S ( )

(16.65)

The factor 4 is made up of a factor of 2 accounting for the change in frequency units and a factor
of 2 accounting for the consideration of positive frequencies only, instead of both positive and
negative frequencies for an even function of frequency.

16.9.7 Cross-correlation Function


It is sometimes necessary to determine the spectral density or autocorrelation function of a quantity
which is a combination of a number of different randomly varying quantity, for each of which the
spectral density or autocorrelation is already known. We may be interested in for example in the
relative motion of two points or the motion of the same point in two orthogonal directions as in rotors.
It is therefore necessary that we be able to determine the spectral density and autocorrelation function
of a sum in terms of those of its components.
Consider a randomly varying quantity s (t ) made up of two quantities x(t ) and y (t ) so that
s(t ) = x(t ) + y (t ) . The autocorrelation function of the combined signal Rs ( f ) can be expressed as

999
Rs ( ) = E [ s (t ) s (t + ) ] = E [{x(t ) + y (t )}{x(t + ) + y (t + )}]

= E [ x(t ) x(t + ) + x(t ) y (t + ) + y (t ) x(t + ) + y (t ) y (t + ) ]

= E [ x(t ) x(t + ) ] + E [ x(t ) y (t + )] + E [ y (t ) x(t + )] + E [ y (t ) y (t + )]

(16.66)

= Rxx ( ) + Rxy ( ) + Ryx ( ) + Ryy ( )


where Rxx ( ) and Ryy ( ) are autocorrelation functions, and Rxy ( ) and Rxy ( ) are crossautocorrelation functions. Cross-correlations are zero if two signals x(t ) and y (t ) are uncorrelated
(for the case of rotor displacements in two orthogonal directions since these displacements are due to
a common unbalance force so they will be correlated, however, two mode shapes of a single system
will be uncorrelated since mode shapes are independent displacement vectors). For stationary process,
we have
Rxy ( ) = E[ x(t ) y (t + )] = E[ x(t ) y (t )] = Ryx ( )

(16.67)

16.9.8 Coherence Function


The coherence function (or coherency squared function) is defined as

(f)=
2
xy

S xy ( f )

S xx ( f ) S yy ( f )

(16.68)

and stratifies for all f


0 xy2 1

(16.69)

16.10 Protection against spurious signals


When studying vibration signals it is important to be sure that the signal under consideration
represents vibration, which is actually present. In some situations it is possible for the electrical signal
received by the measuring system to indicate the presence of vibration that does not in fact exist.
Alternatively, a spurious signal may be superimposed on a genuine vibration signal and thereby upset
its measurement. The two most common causes of the spurious signals are noise and runout.

16.10.1 Electrical noise


Electrical noise may arise from a number of different sources: (i) random electron motion (ii) local
magnetic fields arcing and (iii) earth loop faults.
Noise created by the random motion of electrons is known as thermal noise. It is generally only of
the order of V , however, and is not normally significant in measurements of rotating machinery

1000
vibrations. The noise set up by local magnetic fields is usually of the order of mV and is more
significant. It is set up as a consequence of magnetic fields in nearby electrical apparatus introducing
noise signals in electrical leads conveying the signal that is to be measured. Some protection against
this type of noise-generating mechanism can be obtained by screening the leads with a highconductivity material, which is earthed. A similar effect can also arise when one piece of
instrumentation is sited close to another, because the magnetic fields inside one can induce signals in
the circuit of the other. In this instant the entire instrument should be screened by placing it inside a
metal case. Another source of electromagnetic radiation which can induce noise signals is that emitted
as a consequence of electrical arcing in switches and commutators. This type of noise-generating
mechanism can also be protected against by screening as described above. Earth loop noise can occur
when there are too many earth connections in the instrumentation.

16.10.2 Runout
Proximity transducers used to monitor rotating machinery vibrations depend for their operation on a
change in transducer reactance. Changes in transducer reactance may be present as a consequence of
either mechanical or electrical runout. Mechanical runout is present when the shaft section being
monitored by the transducer is eccentric to the axis of rotation, or has significant surface undulations,
alternatively, the shaft may be bent. Each of these conditions results in motion of the shaft surface
towards or away from the transducer tip when the shaft is rotated, motion which is not caused by shaft
vibration. To remove mechanical runout the shaft must be re-machined; if the shaft is bent it must be
straightness prior to re-machining. Electrical runout is present when, around the shaft circumference,
there are variations in the permeability of the shaft material to the electrical field set up by the
transducer. This might be caused by residual magnetic fields in the shaft surface (due for example to
particular methods of manufacture or non-destructive testing), by material inhomogeneity or by local
residual stresses. Residual magnetic fields can be removed by degaussing, if residual magnetism is not
the cause of the electrical runout then special treatments such as burnishing, micropeening, or
electroplating may be necessary. The amount of mechanical runout present can be determined by
mounting a dial test indicator next to the transducer, and noting the variation in reading when the shaft
is rotated slowly. The electrical runout present may be determined by noting the transducer reading
when the shaft is rotated at low speed (sufficient to obtain a reading on the measuring instrument but
not so high as to cause the shaft to vibrate). The subtraction of the mechanical runout vector from the
measured vector then gives the electrical runout vector. All runout vectors should be removed from
signal measurements to enable true vibration vectors to be recorded. Such nulling is carried out by
sebstarcting the total residual runout vector from the measured vector, as indicated in Figure 16.43.

1001

Measured runout vector

Measured vibration vector

Subtracting the runout vector

Actual runout vector

Figure 16.43 Nullifying of runouts from measured signal

16.10.3 Removing runout from a vibration signal


Some types of tracking filter enable runout vectors to be automatically removed from the incoming
signal, so that the true vibration vector is displaced immediately. This facility can also be programmed
into data logging equipment. The nulling is also carried out between critical speeds when producing a
Bode plot, as discussed earlier, to obtain information about phase for use in balancing. In this case it is
not only the runout that is removed from the incoming signal, but also components of vibration
associated mainly with modes whose characteristics frequencies are lower than the frequency at which
the nulling is being carried out.
Run-out correction
The rotor system has an initial run-out error mainly due to surface irregularities on the shaft. This runout has to be subtracted from the response obtained at the bearings and the steps taken are enumerated
below
1. The rotor is run at a slow speed (for example 240 rpm) and the response at the bearings is
recorded.
2. These responses are split along the time scale for a single time period of rotation into an even
number of divisions, n.
3. The responses taken at higher speeds during the course of the experiment are also split along
the time scale for a single time period of rotation into the same number of divisions, n.
4. On subtracting the slow response data obtained in step 2 from response of step 3, the run-out
error is removed from the responses.
Figure 16.44 shows the procedure for run-out correction

1002

n divisions

Time
Slow speed response

High speed response

Final response

Figure 16.44 Run-out correction

16.10.4 Electronic differentiation and integration


In many instances the vibration signal is to be differentiated or integrated prior to being measured. For
example, many signal conditioning devices enable accelerometer output signals to be integrated twice
to provide a signal which is proportional to displacement, or alternatively they might differentiate a
velocity transducer output signal to provide a measurement of acceleration. Generally, passive
electronic differentiators and integrators results in a loss of signal accuracy of the order of several per
cent because their mathematical foundation involves an approximation. Active devices, requiring a
power source for internal amplifiers, are more accurate but are also more expensive. It is particularly
important to guard against spurious signals when differentiation or integration is necessary. Since the
differentiation requires that the signal is multiplied by its frequency, so that if the spurious signal is of
a higher frequency than the vibration signal then after differentiation the spurious signal may appear
to be far more significant. A similar effect occurs with integration when the spurious signal is of a
lower frequency than the vibration signal.
During the last few years, an emerging interest has been reported on modelling rotating machines
signals as cyclostationary. Some noteworthy precursory works have been concerned with the vibration
signals of gears (Capdessus, et al, 2000; Bouillaut et al., 2001), rolling element bearings (McCormick,
Nandi, 1998; Randall et al., 2001) and reciprocating engines (Antoni et al., 2002). It is very likely
that developments in this area will keep on growing due to the promising successes of these first
applications. Indeed, it is the aim of this paper to demonstrate that virtually any rotating machine
signal can be regarded as being cyclostationary provided the right point of view is adopted.
Cyclostationarity is a property that characterises stochastic processes whose statistical properties
periodically vary with respect to some generic variable. By definition, this embodies a class of nonstationary stochastic processes, with stationary and deterministic periodic processes as special cases.
Specifically due to this generality, cyclostationarity is so nicely suited to rotating machine signals.
Historically, the paradigm of cyclostationarity was first introduced in communication engineering
where the inherent periodicity of the signal statistics is set by the symbol duration (Bennett, 1958). It

1003
is only very recently that it was recognised cyclostationary processes fit the properties of most
rotating machine signals and it can thus be used to devise new processing tools.

Final Remarks
To conclude in the present chapter, various kinds of graphical display used in depicting vibration
responses from rotating machineries are presented, e.g. Bode plots, Nyquist plots, Cascade or
waterfall plots, orbit plots, Campbell diagram, etc. Various aspect in vibration signal measurements
are looked into like digitization error, minimum sampling rate (Nyquist frequency), aliasing effects,
leakage error. Means and ways to reduce these errors during measurements are described in detail like
windowing, tuning of sampling rate, etc. Various statistical properties of random signals are described
like the ensemble and temporal averaging, probability distribution function, probability density, autocorrelation and cross-correlation, coherence, covariance, power spectrum, etc. Various means of
protecting of actual signals from noising signals are looked into due to mechanical and electrical
faults in measuring equipments and test rigs.

1004
References
Antoni, J, Danire, J, Guillet, F, 2002, Effective vibration analysis of IC engines using
cyclostationarity. Part I: a methodology for condition monitoring, Journal of Sound and
Vibration 257(5), 815837.
Bendat, JS, and Piersol, A. G., 2010, Measurement and Analysis of Random Data, John Wiley, New
York.
Bennett, WR, 1958, Statistics of regenerative digital transmission, Bell Systems Technical Journal 37,
15011542.
Bouillaut, L, Sidahmed, M, 2001, Cyclostationary approach and bilinear approach: comparison,
applications to early diagnosis for helicopter gearbox and classification method based on HOCS,
Mechanical Systems and Signal Processing 15(5), 923943.
Capdessus, C, Sidahmed, M, Lacoume, JL, Cyclostationary processes: application in gear faults early
diagnosis, Mechanical Systems and Signal Processing 14(3), 371385.
Churchill, RV, 1941, Fourier Series and Boundary Value Problems, McGraw-Hill, New York.
Fourier, J, 1822, translated by Alexander Freeman (published 1822, translated 1878, re-released
2003). The Analytical Theory of Heat. Dover Publications. ISBN 0-486-49531-0. 2003
unabridged republication of the 1878 English translation by Alexander Freeman of Fourier'
s
work Thorie Analytique de la Chaleur, originally published in 1822.
McCormick, AC, Nandi, AC, 1998, Cyclostationarity in rotating machine vibrations, Mechanical
Systems and Signal Processing 12(2), 225242.
Newland DE, 1991, Random Vibration and Spectral Analysis. Longman, Essex, UK.
Randall, RB, Antoni, J, Chobsaard, S, 2001, The relationship between spectral correlation and
envelope analysis for cyclostationary machine signals, application to ball bearing diagnostics,
Mechanical Systems and System Processing 15(5), 945962.

1005
Exercise Problems
Exercise 16.1 A vibration signal is measured with a sampling interval of 0.002 sec. What is maximum
frequency of the signal, which will be measured without aliasing effects? If you would like to measure
the signals with the maximum frequency three times what you obtained in the first step, what would
be the sampling interval you would like to choose?

Exercise 16.2 In a particular measurement the following time domain data were measured: 0, 1, 1, 0,
1, 1, 0. The total time duration of the measurement was 0.06 sec. Obtain the DFT of the signal, and
plot the real and imaginary parts of the same with respect to the frequency.

Exercise 16.3 Choose a single correct answer from the following multiple choice questions
(i) During the numerical integration of a noisy signal
(a) the high frequency noise will get amplify

(b) the low frequency noise will get amplify

(c) no effect will be there on the noise

(d) the noise will get removed

(ii) For a signal which contains 1 kHz, 3 kHz and 6 kHz frequency signals, if the signal is sampled
with a frequency of 8 kHz, what are the frequency (in kHz) which all will be observed in the sampled
signal

(a) 1, 2, 3

(b) 1, 3, 6

(c) 1, 3, 4

(d) 1, 3

(iii) For a vibration signal the maximum frequency of interest is 1 kHz. What should be the sampling
interval, t, to capture the correct signal?
(A) 0.0005 sec

(B) 0.0001 sec

(C) 0.005 sec

(D) 0.001 sec

(iv) Let the vibration frequency be and the sampling frequency be s. Then aliasing effect would
not occur, when we have

s < 2

(B) s > 2

(C) s =

(D) s <

(v) In signal processing Window Functions are used for


(A) To avoid the aliasing effect

(B) to avoid the leakage error

1006
(C) to facilitate faster DFT

(D) to facilitate noise removal

(vi) In signal processing of a time domain signal to frequency domain signal by DFT, the leakage
error comes due to
(A) perfect synchronisation of captured signal sampling time with its actual time period

(B) non-synchronisation of captured signal sampling time with its actual time period
(C) addition of external noise in the actual vibration signal
(D) aliasing effect
(vii) The complex-DFT is used in rotating machinery signals to find additional information of
(A) amplitude

(B) phase

(C) frequency

(D) direction of rotation

(viii) The basic assumption of the discrete Fourier transform (DFT) is that
(A) The vibration signal must be periodic signal

(B) The captured vibration signal length is the time period of the signal
(C) The captured vibration signal length must be of the infinite time period
(D) It can not be applied to non-periodic vibration signals
(ix) The Fast Fourier Transform (FFT) is
(A) An algorithm to help faster transformation of vibration signal to frequency domain
(B) Theoretically it is same as DFT
(C) Helps in visualing real time spectrum of an vibration signal

(D) All the above cases (A), (B) & (C)

----------------*********----------------

You might also like