You are on page 1of 5

Statistical Process Control (SPC)

Statistical Process Control (SPC) is a system for monitoring, controlling, and improving a
process through statistical analysis. It has many aspects, from control charting to process
capability studies and improvement. Nonetheless, the over-all SPC system of a company may be
broken down into four basic steps: 1) measuring the process; 2) eliminating variances within the
process to make it consistent; 3) monitoring the process; and 4) improving the process. This four-
step cycle may be employed over and over again for continuous improvement.

Bulk of SPC concepts in use today were developed based on the premise that the process
parameter being controlled follows a normal distribution. Any SPC practitioner must be aware
that the parameter must first be confirmed to be normal before being subjected to analysis
concepts based on normal behavior. Thus, any discussion on SPC must be preceded by a
discussion of what a normal distribution is.

The Normal Distribution

The normal distribution (see Fig. 1), normal curve, or bell-shaped curve, is probably the most
recognized and most widely-used statistical distribution. The reason for this is that many
physical, biological, and social parameters obey the normal distribution. Such parameters are
then said to behave 'normally' or, more simply, are said to be 'normal.' The semiconductor
industry has many processes that output data or results that comprise a normal distribution. As
such, it is important for every process engineer to have a firm grasp of what a normal distribution
is.

Aside from the fact that the normal distribution is frequently encountered in our day-to-day lives,
the mathematics governing normal behavior are fairly simple. In fact, only two parameters are
needed to describe a normal distribution, namely, the mean or its center, and the standard
deviation (also known as sigma) or its variability. Knowing both parameters is equivalent to
knowing how the distribution looks like.

The normal distribution is bell-shaped, i.e., it peaks at the center and tapers off outwardly while
remaining symmetrical with respect to the center. To illustrate this in more tangible terms,
imagine taking down the height of every student in a randomly selected Grade 5 class and
plotting the measurements on a chart whose x-axis corresponds to the height of the student and
whose y-axis corresponds to the number of students.

Figure 1. The Normal Distribution

What is expected to emerge from this exercise is a normal curve, wherein a big slice of the
student population will have a height that is somewhere in the middle of the distribution, say 57-
59 inches tall. The number of students belonging to other height groups will be less than the
number of students in the 57"-59" category .
In fact, the number of students decreases at a calculable rate as the height group moves further
away from the center. Eventually you might find only one shortest student at, say, 48", and one
tallest student who probably stands at 66". Lastly, plotting the number of the students falling
under different height ranges of equal intervals will result in a bell-shaped curve. Such a plot is
called a histogram, a simple example of which is shown in Figure 2.

Figure 2. Example of a histogram of heights of students in a Grade 5


class; the y-axis corresponds to the number of students per category

What's notable about normal distributions is that regardless of their standard deviation value, the
% of data falling under a given number of standard deviations is constant. For example, say that
the standard deviation of process 1 is 100, and the standard deviation of process 2 is 200.
Process 1 and Process 2 will have different data distribution shapes (Process 1 being more
stable), but for both processes, 66% of the data under the normal curve will fall within +/- one (1)
standard deviation from the mean of the distribution (i.e., between {mean - 1 sigma} and {mean +
1 sigma}), and 37% of the data will be outside it. Table 1 shows the percentages of data falling
under different numbers of sigma.

Table 1. % Data Falling Under Different Numbers of +/- Sigma


# of Sigma's % of Data Covered % of Data Outside
+/- 1 Sigma 66% 37%
+/- 2 Sigmas 95% 5%
+/- 3 Sigmas 99.73% 0.27%
+/- 4 Sigmas 99.9936% 0.0063%
+/- 5 Sigmas 99.99995% 0.00005%

Skewed Distributions

Perfectly normal curves are hard to come by with finite samples or data. Thus, some data
distributions that are theoretically normal may not appear to be one once the data are plotted,
i.e., the mean may not be at the center of the distribution or there may be slight non-symmetry. If
a normal distribution appears to be 'heavy' or leaning towards the right side of the distribution, it
is said to be skewed to the left. A normal distribution that's leaning to the left is said to be
skewed to the right.

Many response parameters encountered in the semiconductor industry behave normally, which is
why statistical process control has found its way extensively into this industry. The objective of
SPC is to produce data distributions that are stable, predictable, and well within the specified
limits for the parameter being controlled.
In relation to the preceding discussions, this is equivalent to achieving data distributions that are
centered between the specified limits, and as narrow as possible. Good centering between limits
and negligible variation translates to parameters that are always within specifications, which is
the true essence of process control.

Control Charting

It is often said that you can not control something that you do not measure. Thus, every engineer
setting up a new process must have a clear idea of how the performance of this new process is
to be measured. Since every process needs to satisfy customer requirements, process output
parameters for measurement and monitoring are generally based on customer specifications.
Industry-accepted specifications are also followed in selecting process parameters for
monitoring.

Control charting is a widely-used tool for process monitoring in the semiconductor industry. It
employs control charts (see Fig. 3), which are simply plots of the process output data over time.
Before a control chart may be used, the process engineer must first ensure that the process to be
monitored is normal and stable.

A process may have several control charts - one for each of its major output parameters. A new
control chart must have at least the following: the properly labeled x- and y-axes, lines showing
the lower and upper specification limits for the parameter being monitored, and a line showing
the center or target of these specifications. Once a control chart has been set up, the operator
must diligently plot the output data at predefined intervals.

After 30 data points have been collected on the chart (may be less if measurement intervals are
long), the upper and lower control limits of the process may already be computed. Control limits
define the boundaries of the normal behavior of the process. Their values depend only on the
output data generated by the process in the immediate past. Control limits are therefore
independent of specification limits. However, both sets of limits are used in the practice of SPC,
although in different ways.

The lower control limit LCL and the upper control limit UCL of a process may be calculated from
the mean and standard deviation (or sigma) of the plotted data as follows:

LCL = Mean - (3 x Sigma);


UCL = Mean + (3 x Sigma).

Thus, the span from the LCL of a process to its UCL is 6 sigma. The probability of getting points
outside this +/- 3 sigma range is already very low (see Table 1). Getting a measurement outside
this range should therefore warn an engineer that something abnormal is happening, i.e., the
process may be going out of control. This is the reason why these boundaries are known as
'control limits.'

Once the control limits have been included on the control charts (also in the form of horizontal
lines like the specification limits), the operator can start using the chart visually to detect
anomalous trends in the process that she would need to notify the engineer about.
Figure 3. Example of a control chart showing data that are
slightly off-centered, but nonetheless in control and within specs

For instance, any measurement outside the control limits is an automatic cause for alarm,
because the probability of getting such a measurement is low. Four (4) or more consecutively
increasing or decreasing points form a trend that is not normal, and therefore deserves attention.
Six (6) consecutive points on one side of the mean also deserve investigation. When such
abnormalities are observed, the process owner must take an action to bring the process back to
its normal behavior.

Control limits must be recomputed regularly (say, every quarter), to ensure that the control limits
being used by the operator are reflective of the current process behavior.

Read more about: Control Charting.

The Process Capability Indices

Being able to monitor a process for out-of-control situations is one thing; knowing how a process
actually performs is another. Eyeballing the centering and shape of a data distribution can give
us quick, useful information on how the corresponding process behaves, but it is not very helpful
in quantifying the process' actual and potential performance. It is for this reason that statisticians
have come up with methods for expressing the behavior or capability of process distributions in
terms of single numbers known as process capability indices.

Process capability refers to the ability of a process to meet customer requirements or


specification limits, i.e., how consistent its output is in being within its lower and upper spec
limits. A process capability index should therefore be able to indicate how well the process can
meet its specs.

The most basic process capability index is known as the simple process capability index,
denoted by 'Cp'. Cp quantifies the stability of a process, i.e., the consistency of its output. As
mentioned earlier, the process capability indices discussed here presume the normality of the
process. As such, the inconsistency of the process may be measured in terms of the standard
deviation or sigma of the output data of the process. This is what Cp does - it uses the sigma to
quantify the variation of a process, and compares it against the distance between the upper spec
limit (USL) and lower spec limit (LSL) of the process. In mathematical form: Cp = (USL - LSL) /
(6 x Sigma).

The quantity (USL - LSL) is basically the range of output that the process must meet, while 6
sigma corresponds to +/- 3 sigma from the mean, or 99.73% of all the process output data. The
smaller the value of 6 sigma, the narrower the process output distribution is, denoting higher
stability. Thus, Cp increases as process stability increases. Thus, a process needs a Cp > 1 to
ensure that it is narrow enough to meet the spec range 99.73% of the time.
Although Cp indicates the stability of a process, it has one major drawback that makes it almost
useless in the semiconductor industry. It does not consider the centering of the process
distribution within the spec limits. A process with a Cp of 100 may be very stable, with all its
output data very close to each other, but it may also be out-of-spec at all times, i.e., if it is
centered outside the spec limits!

This weakness of Cp is addressed by another process capability index, Cpk. Cpk measures how
centered the output of the process is between its lower and upper limits, as well as how variable
the output is. Cpk is expressed as the ratio of how far the mean of the output data is from the
closer spec limit (the centering of the process) to three times their standard deviation (the
process variability).

CPL = (mean - LSL) / (3 sigma) : process capability index for single-sided (lower) spec limit
CPU = (USL - mean) / (3 sigma) : process capability index for single-sided (upper) spec limit
Cpk = min{CPL,CPL} : process capability index for two-sided spec limits

What these formulae mean is this: Cpk is equal to whichever is lower between CPL and CPU. If
the mean of the process data is closer to the lower spec limit LSL, then Cpk = CPL. If the mean of
the process data is closer to the upper spec limit USL, then Cpk = CPU.

An ideal process is one whose output is always dead center between the spec limits, such that
the mean of its output data equals this dead center and the standard deviation is zero. The Cpk
of this ideal process is infinite (so is the Cpk of other processes whose sigma = 0, as long as the
LSL<mean<USL).

The Cpk decreases if one or both of the following occurs: 1) the data become less centered; and
2) the data become more variable (sigma increases). Thus, improving the process capability of a
process entails one or both of: 1) centering the output between limits and 2) decreasing the
variation of the output data.

The essence of SPC, therefore, is being able to recognize whether a low Cpk is due to the mean
of the process or its sigma, and taking the necessary actions to correct the problem, be it
centering of the data or making them less variable. In any process, the actions needed to center
the output data may be different from what needs to be done to make the data less variable.
Knowledge of this basic SPC principle is therefore a necessary weapon in every process
engineer's arsenal.

As of this writing, most semiconductor companies target a Cpk of 1.67 for their processes,
although they would be satisfied to have an actual Cpk of at least 1.33. Everything, of course,
depends on what spec limits the customer imposes on the manufacturer. Still, at the end of the
day it should always be the manufacturer's goal to center their output between these spec limits
as consistently as possible.

You might also like