You are on page 1of 81

CHAPTER 4

Probability
1. Introduction
2. Probability of an Event
3. Methods of Assigning Probability
4. Event Relations and Two Laws of Probability
5. Conditional Probability and Independence
6. Random Sampling from A Finite Population
1. INTRODUCTION

When the data set represents a sample from a statistical


population, its description is only a preliminary part of a
statistical analysis.

Our major goal is to make generalizations or inferences about


the target population on the basis of information obtained from
the sample data.

An acquaintance with the subject of the probability is essential


for understanding the reasoning that leads to such
generalizations.

In everyday conversations, we all use expression of the kind:


-Most likely our team will win this Saturday
- it is unlikely that the weekend will be cold
-I have a 50-50 chance of getting a summer job at the camp
1. INTRODUCTION

The phrases, most likely, probable, quite unlikely, and so


on used qualitatively to indicate the chance that an event will
occur.

Probability, as a subject, provides a means of quantifying


uncertainty.

In general terms, the probability of an event is a numerical


value that gauges how likely it is that the event will occur.

We let probability take values on a scale from 0 to 1 with very


low values indicating extremely unlikely, values close to 1
indicating very likely, and the intermediate values interpreted
accordingly.
1. INTRODUCTION

A full appreciation for the concept of a numerical measure of


uncertainty and its role in statistical inference can be gained
only after the concept has been pursued to a reasonable
extent.

We can, however, preview the role of probability in one kind of


statistical reasoning.

Suppose it has long been observed that in 50% of the cases


a certain type of muscular pain goes away by itself. A
hypnotist claims that her method is effective in relieving
the pain. For experimental evidence, she hypnotizes 15
patients and 12 get relief from the pain. Does this
demonstrate that hypnotism is effective in stopping the pain
?
1. INTRODUCTION

Let us scrutinize the claim from a statistical point of view. If


indeed the method had nothing to offer, there would still be
a 50-50 chance that a patient is cured. Observing 12 cures
out of 15 amounts to obtaining 12 heads in 15 tosses of a
coin. We will see later that the probability of at least 12
heads in 15 tosses of a fair coin is 0.018, indicating that the
event is not likely to happen. Thus, tentatively assuming
the model (or hypothesis) that the method is ineffective, 12
or more cures are very unlikely. Rather than agree that an
unlikely event has occurred, we conclude that the
experimental evidence strongly supports the effectiveness
of the method.

This kind of reasoning, called testing a statistical hypothesis,


will be explored in greater detail later. For now, we will be
concerned with introducing the ideas that lead to assigned
values for probability.
2. PROBABILITY OF AN EVENT

The probability of an event is viewed as a numerical measure


of the chance that the event will occur.

The idea is naturally relevant to situations where the outcome


of an experiment or observation exhibits variation.

Although we have already used the term of experiment and


event. A more specific explanation is now in order.

In the present context, the term experiment is not limited to


the studies conducted in a laboratory.

Rather, it is used in a broad sense to include any operation of


data collection or observation where the outcomes are subject
to variation.
2. PROBABILITY OF AN EVENT

Rolling a die, drawing a card from a shuffled deck, sampling a


number of customers for an opinion survey, and quality
inspection of items from a production line, are just a few
examples.

An experiment is the process of observing a phenomenon


that has variation in its outcome

Before attempting to assign probabilities, it is essential to


consider all the eventualities of the experiment.

Pertinent to their description, we introduce the following


terminologies and explain them through examples.
2. PROBABILITY OF AN EVENT

The sample space associated with an experiment is the


collection of all possible distinct outcomes of the experiment.

Each outcome is called an elementary outcome, a simple


event, or an element of the sample space.

An event is the set of elementary outcomes possessing a


designated feature.

The elementary outcomes, which together comprise the


sample space, constitute the ultimate breakdown of the
potential results of an experiment.

For instance, in rolling a die, the elementary outcomes are the


point 1, 2, 3, 4, 5, and 6, which together constitute the sample
space.
2. PROBABILITY OF AN EVENT
The outcome of football game would be either a win, loss or
tie.

Each time the experiment is performed, one and only one


elementary outcome can occur.

A sample space can be specified by either listing all the


elementary outcomes, using convenient symbols to identify
them, or making a descriptive statement that characterizes the
entire collection.

For general discussions, we denote


the sample space by
the elementary outcomes by e1, e2, e3,
events by A, B, etc

In specific applications, the elementary outcomes may be


given other labels that provide a more vivid identification.
2. PROBABILITY OF AN EVENT

We say that an event A occurs when any one of the elementary


outcomes in A occurs

Example 1.
Toss a coin twice and record the outcomes head (H) or tail (T)
for each toss.

The sample space can be listed as

= {e1, e2, e3, e4}


2. PROBABILITY OF AN EVENT

We say that an event A occurs when any one of the elementary


outcomes in A occurs

Example 1.
Toss a coin twice and record the outcomes head (H) or tail (T)
for each toss.

The order in which the elements of are listed is


inconsequential.

It is the collection that matters.


2. PROBABILITY OF AN EVENT

Consider the event of getting exactly one head and let us it A.


Scanning the above list, we see that only the elements HT (e2)
and TH (e3) satisfy the requirement.

Therefore, the event A has the composition

A = {e2, e3}

which is, of course, a subset of .


2. PROBABILITY OF AN EVENT

The event B of getting no heads at all, consists of the single


element e4, so B = {e4}. That is, B is a simple event as well as
an event.

The term event is a general term that includes simple event.


2. PROBABILITY OF AN EVENT

Example 2.
Suppose a box containing 50 seeds has 42 seeds that are
viable and 8 not viable. One seed will be picked at random and
planted.
Here the two distinct possibilities are either the chosen seed is
viable or not viable. The sample space can therefore be
structured as
= {V, N}
Alternatively, we may view the choice of each seed as a
distinct elementary outcome, and viable or not viable as its
properties. In other words, we can imagine tagging the viable
seeds with the label e1, e2, , e42, and the non-viable ones
with e43, , e50. Since any of the 50 seeds can be selected by
the process of random choice, the sample space will have 50
elementary outcomes:
= {e1, , e42, e43, ., e50}

viable not viable


2. PROBABILITY OF AN EVENT
Example 2.
Let us denote by V the event that chosen seed is viable. Then,
V has the composition
V = {e1, , e42}

The second formulation of the sample space appears to be


unnecessarily complex in view of the simplicity of the first.

Its advantage will surface latter when we discuss the


assignment of probability to event.

Example 2 shows that, in a given problem, the formulation of


the sample space is not necessarily unique.

Convenience in the determination of probability is a more


important criterion than simplicity of brevity of our list.
2. PROBABILITY OF AN EVENT

Both Example 1 and Example 2 illustrate a sample spaces that


have a finite number of elements.

There are also sample spaces with infinitely many elements.

For instance, suppose a gambler at a casino will continue


pulling the handle of a slot machine until he hits the first
jackpot.

The conceivable number of attempts does not have a natural


upper limit so the list never terminates.

That is, = {1, 2, 3, } has an infinite number of elements.

However, we notice that the elements could be arranged one


after another in a sequence.
2. PROBABILITY OF AN EVENT

An infinite sample space where such an arrangement is


possible is called countably infinite.

Another type of infinite sample space is also important.

Suppose a car with a full tank of gasoline is driven until its fuel
runs out and the distance travelled recorded.

Since distance is measured on a continuous scale, any non-


negative number is a possible outcome.
2. PROBABILITY OF AN EVENT

Denoting the distance traveled by d, we can describe this


sample space as
= {d : d 0}
that is, the set of real numbers greater than or equal to zero.
Here the elements of form a continuum and cannot be
arranged in a sequence. This is an example of a continuous
sample space.

To avoid unnecessary complication, we will develop the basic


principles of probability in the context of finite sample spaces.

We first elaborate on the notion of the probability of an event


as a numerical measure of the chance that it will occur.

The most intuitive interpretation of this quantification is to


consider the fraction of times the event would occur in many
repeated trials of the experiment.
2. PROBABILITY OF AN EVENT

The probability of an event is a numerical value that


represents the proportion of times the event is expected to
occur when the experiment is repeated under identical
conditions.

The probability of event A is denoted P(A)

Since a proportion must lie between 0 and 1, the probability of


an event is a number between 0 and 1.
2. PROBABILITY OF AN EVENT

To explore a few other important properties of probability, let


us refer to the experiment in Example 1 of tossing a coin twice.
The event A of getting exactly one head consists of the
elementary outcomes HT (e2) and TH (e3). Consequently, A
occurs if either of this occurs. Because
2. PROBABILITY OF AN EVENT

the number that we assign as P(A) must be the sum of the two
numbers P(HT) and P(TH).
2. PROBABILITY OF AN EVENT

Guided by the example, we state some general properties of


probability.

The probability of an event is the sum of the probability


assigned to all the elementary outcomes contained in the
event.

Next, since the sample space includes all conceivable


outcomes, in every trial of the experiment some element of
must occur.

Viewed as an event, is certain to occur, and therefore its


probability is 1.

The sum of the probabilities of all the elements of must be 1


2. PROBABILITY OF AN EVENT

In summary,

Probability must satisfy:

(i) 0 P(A) 1, for all events A

(ii)

(iii)
2. PROBABILITY OF AN EVENT

We have deduced these basic properties of probability by


reasoning from the definition that the probability of an event is
the proportion of times the event is expected to occur in many
repeated trials of the experiment.
2. METHODS OF ASSIGNING PROBABILITY

An assignment of probabilities to all the events in a sample


space determines a probability model.

In order to be a valid probability model, the probability


assignment must satisfy the properties (i), (ii), and (iii) stated
in the previous section.

Any assignment of numbers P(ei) to the elementary outcomes


will satisfy the three conditions of probability provided these
numbers are nonnegative, and their sum over all the outcomes
ei in is 1.

However, to be of any practical import, the probability assigned


to an event must also be in agreement with the concept of
probability as the proportion of times the event is expected
occur.
3. METHODS OF ASSIGNING PROBABILITY

Here we discuss the implementation of this concept in two


important situations.

1. Equally likely elementary outcomes the uniform


probability model
2. Probability as the long-run relative frequency
3. METHODS OF ASSIGNING PROBABILITY

Here we discuss the implementation of this concept in two


important situations.

1. Equally likely elementary outcomes the uniform


probability model
2. Probability as the long-run relative frequency
3. METHODS OF ASSIGNING PROBABILITY

Often symmetry in the experiment ensures that each


elementary out-comes is as likely to occur as any other.

For example, consider the experiment of rolling a perfect die


and recording the top face.

The sample space can be described as


= {e1, e2, e3, e4, e5, e6}
where e1 stands for the elementary outcome of getting the
face 1, and similarly, e2, , e6.

Without actually rolling a die, we can deduce the probabilities.

Because fair die is a symmetric cube, each of its six faces is as


likely to appear as any other.

In other words, each face is expected to occur one-sixth of the


time.
3. METHODS OF ASSIGNING PROBABILITY

The probability assignments should therefore be


P(e1) = P(e2) = = P(e6) = 1/6
and any other assignment would contradict the symmetry
assumption that the die is fair.

We say that rolling a fair die conforms to a uniform


probability model because the total probability 1 is evenly
apportioned to all the elementary outcomes.

What is the probability of getting a number higher than 4 ?


Letting A denote this event, we have the composition A = {e5,
e6}, so
P(A) = P(e5) + P(e6) = 1/6 + 1/6 = 2/6
3. METHODS OF ASSIGNING PROBABILITY

When the elementary outcomes are modeled as equally likely,


we have a uniform probability model. If there are k elementary
outcomes in , each is assigned the probability of 1/k.

An event A consisting of m elementary outcomes, is then


assigned
3. METHODS OF ASSIGNING PROBABILITY

Example 3.
Find the probability of getting exactly one head in two tosses
of a fair coin.

As listed in Example 1, there are four elementary outcomes in


the sample space:

= {HH, HT, TH, TT}

The very concept of a fair coin implies that the four elementary
outcomes in are equally likely. We therefore assign the
probability to each of them.

The event A = [one head] has two elementary outcomes-


namely, TH and HT. Hence, P(A) = 2/4 = 0.5
3. METHODS OF ASSIGNING PROBABILITY

Example 4.
Refer to Example 2. If one seed is sampled at random from a
box that contains 42 viable and 8 non-viable seeds, what is the
probability that the selected seed is not viable ?

The intuitive notion of random selection is that no seed is given


preference over any other.

We deduce that each seed is as likely to be selected as any


other.

Identifying the selection of each seed as an elementary


outcomes, the sample space has the composition
= {e1, , e42, e43, ., e50}

viable not viable


and this specification conforms to a uniform probability model.
3. METHODS OF ASSIGNING PROBABILITY

Each of the 50 elementary outcomes is to be assigned the


probability 1/50.

The event not viable has the composition


[Not viable] = {e43, , e50}
Consequently,
P(not viable) = 8/50 = 0.16.

An alternative reasoning: in the above analysis, the individual


seeds were not, of course, actually tagged. The tagging was
imagined for the purpose of calculating the probabilities.
Alternatively, one may argue that since the selected seed must
be either viable (V) or not viable (N), there are only two
elementary outcomes in the sample space, = {V,N}.
Since the numbers of viable and non-viable seeds are in the
ratio 42:8 in the population, their probabilities must also be in
the same ratio that is
3. METHODS OF ASSIGNING PROBABILITY

While, this second reasoning appears more attractive because


of its simplicity, it runs into considerable complexities when the
experiment involves sampling two or more seeds.
3. METHODS OF ASSIGNING PROBABILITY

Here we discuss the implementation of this concept in two


important situations.

1. Equally likely elementary outcomes the uniform


probability model
2. Probability as the long-run relative frequency
3. METHODS OF ASSIGNING PROBABILITY

Symmetry among the outcomes, when it is present, leads to


uniform probability model. However, the simple model does not
always apply.

In many situations, it is not possible to construct a sample


space where the elementary outcomes are equally likely.

If one corner of a die is cut off, it would be unreasonable to


assume that the faces remain equally likely and the
assignments of probability to various faces can no longer be
made by deductive reasoning.

When speaking of the probability (risk) that a man will die in


his thirties, one may choose to identify the occurrence of death
at each decade or even each year of ages as an elementary
outcomes.
3. METHODS OF ASSIGNING PROBABILITY

However, no sound reasoning can be provided in favor of a


uniform probability model.

In fact, from extensive mortality studies, the demographers


have found considerable disparity in the risk of death at
different age groups.

When the assumption of equally likely elementary outcomes is


not tenable, how do we assess the probability of an event ?

The only recourse is to repeat the experiment many times and


observe the proportion of times the event occurs.
3. METHODS OF ASSIGNING PROBABILITY

Letting N denote the number of repetitions (or trials) of an


experiment, we set

rN(A) = relative frequency of event A in N trials

For instance, let A be the event of getting a six when rolling a


die. If the die is rolled 100 times and six comes up 23 times,
the observed relative frequency of A would be r100(A) = 23/100
=0.23. In another set of 100 tosses, six may come up 18
times, resulting in 18/100 = 0.18 as the relative frequency.
Collecting these two sets together, we have N = 200 trials with
the observed relative frequency
3. METHODS OF ASSIGNING PROBABILITY

Imagine that this process is continued by recording the results


from more and more tosses of the die, and updating the
calculation of relative frequency.

Figure 1 shows a typical plot of the relative frequency rN(A)


versus the number N of trials of the experiment.
3. METHODS OF ASSIGNING PROBABILITY

We see that rN(A) fluctuates as N changes, but the fluctuations


become damped with increasing N, and rN(A) tends to stabilize
as N gets very large.

Two persons separately performing the same experiment N


times are not going to get exactly the same graph of rN(A).
However, the numerical value at which rN(A) stabilizes in the
long-run, will be the same.
3. METHODS OF ASSIGNING PROBABILITY
Let us designate this feature of Figure 1 as the long-run
stability of relative frequency.
We define P(A), the probability of an event A, as the value to
which rN(A) stabilizes with increasing N.

Although we will never know P(A) exactly, it can be estimated


accurately by repeating the experiment many times and
calculating rN(A).

The property of the long-run stabilization of relative


frequencies is based on the findings of experiments in many
fields who have undertaken the strain of studying the behavior
of rN(A) under prolonged repetitions of their experiments.
3. METHODS OF ASSIGNING PROBABILITY
French gamblers, who provided much of the early impetus for
the study of probability, performed experiments tossing die
and coins, drawing cards, and playing other games of chance
thousands and thousands times.

They observed the stabilization property of relative frequency


and applied this knowledge to achieve an understanding of the
uncertainty involved in these games.

Demographers have compiled and studied volumes of mortality


data to examine the relative frequency of the occurrence of
such events as death in particular age groups.

In each context, the relative frequencies were found to


stabilize at specific numerical values as the number of cases
studied increased.
3. METHODS OF ASSIGNING PROBABILITY
Life and accident insurance companies actually depend upon
the stability property of relative frequencies.

We explained in Section 3.1 how a determination of probability


can be based on logical deductions in the confines of a uniform
probability model.

When we assign equal probabilities to the six faces of a die, we


essentially conceptualize the die as a perfectly symmetric and
homogenous cube.

In the real world, we many never know if that is indeed the


case.

However, examination of the relative frequencies after a large


number of tosses, can help us assess whether the idealized
model is tenable.
3. METHODS OF ASSIGNING PROBABILITY

Mendel conducted experiments over a period of years before


deducing the validity of the uniform probability for gene
selection.

As another example of an idealized model, consider the


assignment of probabilities of the day of the week a child will
be born.

We may tentatively assume the simple model that all seven


days of the week are equally likely.

Each day is then assigned the probability 1/7/

If A denotes the event of a birth on the weekend (Saturday or


(Sunday), our model leads to the probability P(A) = 2/7.
3. METHODS OF ASSIGNING PROBABILITY

The plausibility of the uniform probability model can only be


ascertained from an extensive set of birth records.

Table 1 shows the frequency distribution of the number of


births by days of the weeks for all registered births in the USA
in the year 1971.
3. METHODS OF ASSIGNING PROBABILITY

An examination of the relative frequencies in Table 1 reveals a


sizable dip on Saturday and Sunday.

Specifically, the relative frequency of births on the weekend is


0.256, as compared to the probability of 2/7 = 0.286
suggested by the uniform probability model.
3. METHODS OF ASSIGNING PROBABILITY

Because the number of trials is so large, the difference 0.286


0.256 = 0.03 appears real.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Recall that the probability of an event A is the sum of the


probabilities of all the elementary outcomes that are in A.

It often turns out, however, that the event of interest has a


complex structure that requires tedious enumeration of its
elementary outcomes.

On the other hand, this event may be related to other events


that can be handled more easily.

The purpose of this section is to first introduce the three most


basic event relations : complement, union, and
intersection.

These event relations will then motivate some laws of


probability.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

The event operations are conveniently described in graphical


terms.

We first represent the sample space as a collection of points in


a diagram, each identified with a specific elementary outcome.

The geometric pattern of the plotted points is irrelevant.

What is important is that each point be clearly tagged to


indicate which elementary outcome it represents, and to watch
that no elementary outcome is missed or duplicated in the
diagram.

To represent an event A, identify the points that correspond to


the elementary outcomes in A, enclose them in a boundary line
and attach the tag A.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

This representation, called a Venn diagram, is illustrated in


Figure 2.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 5.
Make a Venn diagram for the experiment of tossing a coin
twice, and indicate the following events:
A: Tail at the second toss
B : At least one head

Here the sample space is


= {HH, HT, TH, TT},

and the two events have the


compositions A = {HT, TT),
B = {HH, HT, TH}

Figure 2 shows the Venn


diagram.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 6.
Listed below are the types and ages of four monkeys procured
by a laboratory for a drug trial.

Suppose two monkeys will be selected by lottery and assigned


to an experimental drug. Considering all possible choices of
two monkeys, make a Venn diagram and show the following
events:
A: The selected monkeys are of the same type.
B: The selected monkeys are of the same age.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 6.

Here the elementary outcomes are the possible choices of a


pair of numbers from {1,2,3,4}. These pairs are listed and
labeled as e1, e2, e3, e4, e5, e6 for ease reference:
{1,2} (e1) {2,3} (e4)
{1,3} (e2) {2,4} (e5)
{1,4} (e3) {3,4} (e6)
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 6.

{1,2} (e1) {2,3} (e4)


{1,3} (e2) {2,4} (e5)
{1,4} (e3) {3,4} (e6)

The pair {1,2} has both monkeys of the same type, and so
does the pair {3,4}. Consequently, A = {e1, e6}.

Those with the same ages are {1,3}, {1,4} and {3,4}, so
B = {e2, e3, e6}.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 6.
{1,2} (e1) {2,3} (e4)
{1,3} (e2) {2,4} (e5)
{1,4} (e3) {3,4} (e6)

A = {e1, e6}.
B = {e2, e3, e6}.

Figure 3 shows the Venn diagram.


4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

We now proceed to define the three basic event operations and


introduce the corresponding symbols.

The complement of an event A, denoted by , is the set of


all elementary outcomes that are not in A. The occurrence of
means that A does not occur.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

The union of two events A and B, denoted by A U B, is the set


of all elementary outcomes that are in A, in B, or in both. The
occurrence of A U B means that either A or B or both occur.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

The intersection of two events A and B, denoted by AB, is the


set of all elementary outcomes that are in A and in B. The
occurrence of AB means that both A and B occur.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Note that A U B is a larger set containing A as well as B while


AB is the common part of the sets A and B.

Also it is evident from the definitions that A U B and B U A


represent the same event, while AB and BA are both
expression for the intersection of A and B.

The operation of union and intersection can be extended to


more than two events.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Note that A U B is a larger set containing A as well as B while


AB is the common part of the sets A and B.

Also it is evident from the definitions that A U B and B U A


represent the same event, while AB and BA are both
expression for the intersection of A and B.

The operation of union and intersection can be extended to


more than two events.

For instance, A U B U C stands for the set of all points that are
in at least one of A, B, and C; while ABC represents the
simultaneous occurrence of all three events.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Two events A and B are called incompatible or mutually


exclusive if their intersection AB is empty. Because
incompatible events have no elementary outcomes in common,
they cannot occur simultaneously.

Example 7.
Refer to the experiment in Example 6 of selecting two monkeys
out of four. Let A = [same type], B = [same age], and C =
[different types]. Give the compositions of the events

C, , A U B, AB, BC
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

{1,2} (e1) {2,3} (e4)


{1,3} (e2) {2,4} (e5)
{1,4} (e3) {3,4} (e6)

Refer to Example 6,
A = {e1, e6}
B = {e2, e3, e6}

The pairs consisting of different types are {1,3}, {1,4}, {2,3} and
{2,4}, so C = {e2, e3, e4, e5}.

The event is the same as the event C.

Employing the definitions of union and intersection, we obtain

A U B = {e1, e2, e3, e6} AB = {e6} BC = {e2, e3}


4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Let us now examine how probabilities behave as the operations


of complementation, union, and intersection are applied to
events.

It would be worthwhile for the reader to review the properties


of probability listed :
Probability must satisfy:

(i) 0 P(A) 1, for all events A

(ii)

(iii)
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

First, let us examine how P( ) is related to P(A).

The sum P(A) + P( ) is the sum of probabilities of all


elementary outcomes that are in A plus the sum of
probabilities of elementary outcomes not in A.

Together, the two set comprise and we must have P() = 1.


Consequently, P(A) + P( ) = 1, and we arrive at the following
law:

Law of Complementation: P(A) = 1 P( )

This law or formula is useful in calculating P(A) when is of a


simpler form than A so that P( ) is easier to calculate.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Turning to the operation of union, recall that A U B is


composed of points (or elementary outcomes) that are in A, in
B, or in both A and B.

Consequently, P(A U B) is the sum the probabilities assigned


these elementary outcomes, each probability taken just once.

Now, the sum P(A) + P(B) includes contributions from all these
points, but it double counts those in the region AB (see the
figure A U B).

To adjust for this double counting, we must therefore subtract


P(AB) from P(A) + P(B).

This results in the following law:

Addition Law: P(A U B) = P(A) + P(B) P(AB)


4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Addition Law: P(A U B) = P(A) + P(B) P(AB)

If the event A and B are incompatible, their intersection AB is


empty, so P(AB) = 0, and we obtain

For incompatible events: P(A U B) = P(A) + P(B)

The addition law expresses the probability of a larger event A


U B in term of the probabilities of the smaller events A, B and
AB.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 8.
A child is presented with three word-association problems.
With each problem two answers are suggested one is correct
and the other wrong. If the child has no understanding of the
words whatsoever, and answers the problems by guessing,
what is the probability of getting at least one correct answer ?

Let us denote a correct answer by C and a wrong answer by W.


The elementary outcomes can be conveniently enumerated by
means of three diagram.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

There are 8 elementary outcomes in the sample space and,


since they are equally likely, each has probability 1/8.

Let A denote the event of getting at least one correct answer.

Scanning our list, we see that A contains 7 elementary


outcomes, all except WWW.

Our direct calculation yields P(A) = 7/8.


4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Now, let us see how this probability calculation could be


considerably simplified.

First, making a complete list of the sample space is not


necessary.

Since the elementary outcomes are equally likely, we need


only determine that there are a total of 8 elements in .
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

How can we obtain this count without making a list ?

Note that an outcome is represented by three letters. There


are two choices for each letters namely, C or W.

We then have 2 x 2 x 2 = 8 ways of filling the three slots.

The tree diagram explains this multiplication rule of counting.


4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Evidently, the event A contains many elementary outcomes.

On the other hand, is the event of getting all answers wrong.

It consists of the single elementary outcome WWW, so P( ) =


1/8.

According to the law of complementation.

P(A) = 1 P( )
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 9.
Refer to Example 6 where two monkeys are selected from four
by lottery. What is the probability that the selected monkeys
are either of the same type or the same age ?

In Example 6, we already
enumerated the 6 elementary
outcomes that comprise the sample
space.

The lottery selection makes all choices equally likely and the
uniform probability model applies.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 9.
Refer to Example 6 where two monkeys are selected from four
by lottery. What is the probability that the selected monkeys
are either of the same type or the same age ?

{1,2} (e1) {2,3} (e4)


{1,3} (e2) {2,4} (e5)
{1,4} (e3) {3,4} (e6)

The two events of interest are


A = [same type] = {e1, e6}
B = [same age] = {e2, e3, e6}
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 9.
Refer to Example 6 where two monkeys are selected from four
by lottery. What is the probability that the selected monkeys
are either of the same type or the same age ?

The two events of interest are


A = [same type] = {e1, e6}
B = [same age] = {e2, e3, e6}

Because A consists of 2 elementary outcomes and B consists of


3,
P(A) = 2/6 and P(B) = 3/6
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 9.
Refer to Example 6 where two monkeys are selected from four
by lottery. What is the probability that the selected monkeys
are either of the same type or the same age ?

Here, we are to calculate P(A U B).


To employ the addition law, we
also need to calculated P(AB)

From Figure 3, we see AB = {e6},


so P(AB) = 1/6.
4. EVENT RELATIONS AND TWO LAWS OF PROBABILITY

Example 9.
Refer to Example 6 where two monkeys are selected from four
by lottery. What is the probability that the selected monkeys
are either of the same type or the same age ?

P(A) = 2/6
P(B) = 3/6
P(AB) = 1/6.

Therefore,
P (A U B) = P(A) + P(B) P(AB)
= 2/6 + 3/6 1/6
= 4/6
= 2/3

which is confirmed by the observation that A U B = {e1, e2, e3,


e6} indeed has four outcomes.
5. CONDITIONAL PROBABILITY AND INDEPENDENCE

The probability of an event A must often be modified after


information is obtained as to whether or not a related event B
has taken place.

Information about some aspect of the experimental results


may therefore necessitate a revision of the probability of an
event concerning some other aspect of the results.

The revised probability of A when it is known that B has


occurred is called the conditional probability of A given B
and is denoted by . .

To illustrate how such a modification is made, we consider an


example that will lead us to the formula for conditional
probability.
5. CONDITIONAL PROBABILITY AND INDEPENDENCE

Example 10.
A group of executive is classified according to the status body
weight and incidence of hypertension. The proportions in the
various categories appear in Table 2.

(a) What is the probability that a person selected at random


from this group will have hypertension ?
(b) A person, selected at random from this group, is found to
be overweight. What is the probability that this person is also
hypertensive ?
5. CONDITIONAL PROBABILITY AND INDEPENDENCE

Let A denote the event that a person is hypertensive, and let B


denote the event that a person is overweight.

(a) Since 20% of the group is hypertensive and the individual


is selected random from this group, we conclude that
P(A) = 0.2. This is the unconditional probability of A.
5. CONDITIONAL PROBABILITY AND INDEPENDENCE

(b) When we are given information that the selected person is


overweight, the categories in the second and third columns
of Table 2 are no relevant of this person. The first column
shows that among the subgroup of overweight persons, the
proportion having hypertension is 0.10/0.25. Therefore,
given the information that the person is in this subgroup,
the probability that he or she is hypertensive is
5. CONDITIONAL PROBABILITY AND INDEPENDENCE

Noting that

You might also like