You are on page 1of 157

LOGIC:

An
INTRODUCTION
to the
FORMAL STUDY
of
REASONING

JOHN L. POLLOCK
University of Arizona
1
The Formal Study
of Reasoning

1. Problem Solving and Reasoning


Human beings are unique in their ability to acquire sophisticated knowl-
edge about the way the world works and then use that knowledge to achieve
their goals. In using their knowledge and figuring out how to achieve a
goal, they are engaging in problem solving. Nonhuman animals also engage
in problem solving, but they are quite limited in the complexity of the
problems they can solve. Although my cat may be able to solve the problem
of catching a bird by climbing a tree, she will never figure out how to send
a rocket to the moon.
Consider a simple instance of problem solving. This is something
chimpanzees can do. The chimp is put in a cage with a large box, and
bananas are hanging from the ceiling. The chimp wants the bananas. The
problem is how to get them. The chimp solves the problem by moving the
box under the bananas and climbing on top of it, from which it can reach
the bananas. We can only conjecture how chimps solve this problem, but
humans solve the same problem by reasoning. Using our knowledge of the
world, we observe that the bananas are in a certain location. We know that
we can push the box under the bananas, we know that we can climb on top
of the box, and we know that if we were standing on top of the box and it
was beneath the bananas then we could reach the bananas. Finally, we
know that if we could reach the bananas, we could retrieve them. So we
reason that if we push the box beneath the bananas and climb on top of it,
then we can retrieve the bananas. Our reasoning is as follows:

1. If I push the box beneath the bananas, then it will be beneath the
bananas.
2. If the box is beneath the bananas and I climb on top of it, then I will
be able to reach the bananas.
3. If I am able to reach the bananas, then I can retrieve them.

Therefore, if I push the box beneath the bananas and climb on top of it,
then I can retrieve the bananas.

The reasoning we employed in solving this problem can be generalized.


It has a general structure that can be revealed by abbreviating “I push the
box beneath the bananas” as P, “the box will be beneath the bananas” as B,
“I climb on top of the box” as C, “I am able to reach the bananas” as A, and
“I can retrieve the bananas” as R:
THE FORMAL STUDY OF REASONING 4

1. If P, then B.
2. If B and C, then A.
3. If A, then R.

Therefore, if P and C, then R.

What should be noticed is that this reasoning would be equally good reasoning
if we let P, B, C, A, and R abbreviate anything else. It is the general
structure of the reasoning that makes it good reasoning.
To solve more complex problems, we may have to engage in more
complex reasoning. For instance, if I want to see a movie, and the only way
I can get there is by taxi, I may reason as follows:

1. If I am at the movie theatre and I buy a ticket, I can see the movie.
2. If I am at the movie theatre and I have seven dollars, I can buy a
ticket.
3. If I am at home and I pay a cab driver ten dollars and ask him to
drive me to the theatre, he will do so.
4. If the cab driver drives me to the theatre, then I will be at the theatre.
5. If I pay the cab driver ten dollars, then I will have ten dollars less
than I did to begin with.

Therefore, if I am at home and have seventeen dollars, I pay a cab driver


ten dollars and ask him to drive me to the theatre, and I buy a ticket, I
can see the movie.

This reasoning is much like the reasoning involved in retrieving the bananas,
except that it also involves some mathematical reasoning.
Solving different kinds of problems may require different kinds of
knowledge about the world and different kinds of reasoning. If my goal is
to send a rocket to the moon, I will have to know a lot about physics and
engineering, and my reasoning will be heavily mathematical. If my goal is
to be elected mayor of Tucson, I will need different kinds of knowledge and
the reasoning will also be different. If my goal is to write a logic text, the
required knowledge and reasoning will be different yet.

2. Rationality, Logic, and Reasoning


One of the traditional topics of philosophy is rational thought. We
want to know how it works. How is it possible for us to acquire sophisticated
knowledge of the world and use that to solve complex problems? Reasoning
is an important part of rational thought, so the study of good reasoning is
part of the study of rationality.

2.1 Varieties of Reasoning


THE FORMAL STUDY OF REASONING 5

We engage in many different kinds of reasoning:


• Some of our reasoning is “perceptual”. If I look at a book on my
desk in full view and in good light and it looks red to me, then I can
reasonably conclude that (probably) it is red. Most of our knowledge
of the world begins with such perceptual reasoning.
• Another kind of reasoning is inductive. In inductive reasoning, we
draw general conclusions from a number of observations of particular
instances. For example, most people would concur that earthworms
don’t have wings. Why do they believe that? Presumably, because
they have seen many earthworms, and none of those they have seen
have had wings. So they reason to the general conclusion that earth-
worms don’t have wings from their beliefs about individual earth-
worms.
• We cannot directly observe the mental states of other people, but we
form beliefs about their mental states by observing their behavior
and comparing it with our own when we are in various mental states.
• We form scientific theories by constructing explanations for observa-
tions. For instance, the atomic theory of matter was originally pro-
pounded to explain diverse physical phenomena such as the peculiar
spectra of light that substances emit when heated.
• We often engage in mathematical reasoning. For example, if Jones
gives me five dollars and Smith gives me seven, I can infer that
together they gave me twelve dollars.
These varieties of reasoning are characterized in part by their subject
matter. Perceptual reasoning has to do with making inferences from percep-
tual input. Inductive reasoning has to do with reasoning about general-
izations. Reasoning about the mental states of others is about other minds.
And so on. The field of philosophy that studies these diverse kinds of
reasoning is epistemology. Epistemologists try to clarify the way each of
these kinds of reasoning works, and explain why they are examples of good
reasoning.

2.2 Formally Good Reasoning


Logic is also about reasoning, but focuses on a different aspect of it.
Sometimes, the goodness of a bit of reasoning has nothing to do with its
subject matter. The analogous reasoning is correct when applied to any
subject matter. We saw an example of this above. No matter what P, B, C,
A, and R might be, the following reasoning is good reasoning:

1. If P, then B.
2. If B and C, then A.
3. If A, then R.

Therefore, if P and C, then R.

It is the form of the reasoning that makes it good rather than its content.
Logic is the study of this phenomenon. Consider another example. The
following is clearly good reasoning:
THE FORMAL STUDY OF REASONING 6

1. All Republicans are faithful to their spouses.


2. No one who is faithful to his or her spouse will ever get into political
trouble.

Therefore, no Republican will ever get into political trouble.

One may doubt the truth of the premises, but the reasoning is impeccable.
That is, given the premises, one can infer the conclusion. The correctness of this
inference, however, has nothing to do with the fact that it is about Republicans,
spouses, and political trouble. The inference has the general form:

1. All A are B.
2. No B is C.

Therefore, no A is C

Here is another instance of the same form of reasoning:

1. All Republicans are unfaithful to their spouses.


2. No one who is unfaithful to his or her spouse will ever be elected
President.

Therefore, no Republican will ever be elected President.

This reasoning is equally good, and for the same reason. Again, one might
object to the premises, but given the premises, the conclusion follows.
Reasoning that is good because of the form of the reasoning can be
called formally good reasoning. Logic is the study of formally good reasoning.
In logic we investigate what forms of reasoning generate formally good
reasoning.

2.3 Soundness and Inferential Correctness


The preceding examples illustrate an important distinction. We can
evaluate reasoning in two different ways. Reasoning begins from premises
and draws a conclusion. One way to evaluate the reasoning is by asking
whether the premises it assumes are true. Neither of the above arguments
fares well on this dimension of evaluation. But we can also ask whether,
given the premises, the argument supports the conclusion. On this dimension
of evaluation, both of the preceding arguments are flawless.
Let us say that reasoning is inferentially correct when it is good on the
second dimension. Inferentially correct reasoning is reasoning wherein the
inferences are correct, regardless of whether the premises are true. All of the
reasoning we have discussed so far has been inferentially correct. However,
some of that reasoning has proceeded from false premises. Reasoning is
said to be sound when it is inferentially correct reasoning from true premises.
Sound reasoning is reasoning that survives evaluation on both of the above
dimensions.
Logic does not concern itself with soundness—only with inferential
THE FORMAL STUDY OF REASONING 7

correctness. To know whether reasoning is sound, we must investigate the


truth of the premises, and depending upon what the premises are, that may
take us far afield from philosophy into physics, history, economics, geogra-
phy, or sometimes just common sense. Inferential correctness, on the other
hand, can be investigated without engaging in scientific investigations of
the world. In exploring the concept of inferential correctness, we can
ignore the truth or falsity of the premises and just ask whether the conclusion
would follow if the premises were true.
The terminology “sound argument” is traditional in logic, but it is also
unfortunate. As just noted, we will not have occasion to talk about sound
arguments, so the terminology is not useful. To make matters worse, when
we get further into the study of logic, we will find that the term “sound”
has other uses unrelated to its use in talking about sound arguments, and
those other uses are more important.

3. Arguments and Inferences


Reasoning proceeds by making inferences. We start with some premises,
and end with a conclusion, but typically there will be a number of steps in
between. Consider the following puzzle:

A murder has been committed, and an eyewitness reports that

1. The murderer was either John or Joe,

but they are identical twins so the eyewitness could not tell them
apart. The detectives have discovered the following clues:

2. The murder was committed in Tucson.


3. One of the brothers was in Phoenix at the time of the crime, but the
murderer was in Tucson.
4. John was in Phoenix only if Joe had breakfast with Maria on the
morning of the crime.
5. Joe had a fight with Maria the night before, and would not have had
breakfast with Maria unless they made up.
6. Maria would only make up with Joe if John asked her to.
7. John only asked Maria to make up with Joe if John was in Tucson.

We also know:

8. If someone was in Tucson at a certain time, they were not in Phoenix


at that same time.

Who was the murderer?

We can solve this puzzle by reasoning from the premises in the following
way:
THE FORMAL STUDY OF REASONING 8

9. It follows from (1) that if John wasn’t the murderer, Joe was.
10. It follows from (3) that either John was in Phoenix or Joe was in
Phoenix.
11. Suppose John is not the murderer. Then we can reason as follows:
12. From (11) it follows that Joe is the murderer.
13. From (2) and (12) it follows that Joe was in Tucson.
14. From (8) and (13) it follows that Joe was not in Phoenix.
15. From (3) and (14) it follows that John was in Phoenix.
16. From (4) and (15) it follows that Joe had breakfast with Maria.
17. From (5) and (16) it follows that Joe made up with Maria.
18. From (6) and (17) it follows that John asked Maria to make up
with Joe.
19. From (7) and (18) it follows that John was in Tucson.
20. From (3) and (19) it follows that Joe was in Phoenix, and hence
was not the murderer.
21. From (1) and (20) it follows that John is the murderer.
22. So from the supposition that John is not the murderer, it follows that
he is. Hence that supposition cannot be true, so:

23. John is the murderer.

Consider how this puzzle was solved. We began with premises (1)–(8), and
we ended with the conclusion (23). However, it is not initially obvious that
the conclusion follows from the premises, so we fill the gap with a number
of intermediate inferences each of which is obvious by itself. Stringing
those inferences together allows us to establish the conclusion. This illustrates
that reasoning is essentially a tool whose purpose is to enable us to acquire
new knowledge on the basis of things we already know.
When we represent our reasoning in this way, what we write is an
argument. An argument is a list of statements each of which is either a
premise or inferred from earlier steps of the argument. The individual
steps represent either premises or inferences from previous steps. It is con-
venient to number the steps of the argument, and include an explanation of
how each statement is inferred.

Exercises. Construct arguments solving the following puzzles:

First Puzzle:
When Arthur, Bill, and Carter eat out:
1. Each orders either ham or pork (but not both).
2. If Arthur orders ham, Bill orders pork.
3. Either Arthur or Carter orders ham, but not both.
4. Bill and Carter do not both order pork.
Who could have ordered ham one day and pork another day?
THE FORMAL STUDY OF REASONING 9

Second Puzzle:
Fred knows five women: Ann, Betty, Carol, Deb, and Eve.
1. The women are in two age brackets: three women are under 30, and two
women are over 30.
2. Two women are teachers and the other three women are secretaries.
3. Ann and Carol are in the same age bracket.
4. Deb and Eve are in different age brackets.
5. Betty and Eve have the same occupation.
6. Carol and Deb have different occupations.
7. Of the five women, Fred will marry the teacher over thirty.
8. Each of these women is either a secretary or a teacher but not both.
9. Fred will marry one of these women.
Whom will Fred marry?

4. Two Kinds of Reasoning


4.1 Deductive Reasoning
There is a distinction between two kinds of reasoning. Some reasoning
has the characteristic that if the premises are true then the conclusion could
not fail to be true. For such reasoning, there are no conceivable circumstances
under which the premises are true but the conclusion false. Most of the
arguments we have considered have this characteristic. For example, recall
the argument:

1. All Republicans are faithful to their spouses.


2. No one who is faithful to his or her spouse will ever get into political
trouble.

Therefore, no Republican will ever get into political trouble.

We can certainly imagine circumstances in which the conclusion is false. In


fact, the real world is probably one such circumstance. But we cannot
imagine circumstances in which the conclusion is false while the premises are
true. The only way the conclusion of this argument could be false is by
having one or more premises false as well. Such an argument is deductive.
In a deductive argument, the conclusion follows of necessity from the premises.

4.2 Defeasible Reasoning


Not all good reasoning proceeds in terms of deductive arguments.
Section two briefly surveyed a number of different kinds of arguments that
are important to our being able to acquire knowledge of the world, and
most of those arguments are non-deductive. For example, consider inductive
reasoning. One might reason as follows:
THE FORMAL STUDY OF REASONING 10

I have observed many earthworms, and none have had wings.

Therefore, no earthworms have wings.

It is clear that this is not a deductive argument. The fact that you have not
observed any earthworms with wings does not make it impossible for there
to be some. Europeans once reasoned similarly about swans, concluding
that all swans are white. But it turned out that there are black swans in
Australia.
It is a general characteristic of inductive reasoning that it is not deductive.
In fact, that is the whole point of it. The value of inductive reasoning is that
it allows us to discover general truths about the world. We can never
observe general truths. All we can do is observe particular instances of
them, and infer what the general truths are on the basis of the particular
instances. This is what science is all about.
There is a temptation to suppose that if inductive reasoning is not
deductive, then there is something wrong with it. But that is to mistake
what the purpose of reasoning is. Reasoning is a tool of rationality, and
what it does is lead us to reasonable beliefs. The beliefs must be reasonable,
but they need not be infallible. None of our interesting beliefs about the
world are infallible. Try to think of anything you know that you know on
the basis of premises that deductively guarantee its truth. The only examples
you are apt to find are mathematical examples or simple tautologies like “It
is snowing or it isn’t snowing”. Virtually all of the copious knowledge you
have of the world and in terms of which you successfully solve practical
problems is had on the basis of reasoning that is not deductive.
Good arguments that are not deductive make their conclusions reason-
able, but do not absolutely guarantee their truth. When we employ such
arguments, we must always be prepared to retract the conclusions if subse-
quent discoveries mandate that. Such arguments are said to be defeasible,
because they can be “defeated”. The considerations that defeat a defeasible
inference are called defeaters. For example, consider the earthworm argument.
What might a defeater be? Here is an obvious example:

This is an earthworm and it has wings.

In other words, no matter how many confirming instances we have for a


generalization, a single counter-example suffices to show it false and defeat
the inductive inference. This is how the reasoning supporting the general-
ization “All swans are white” was defeated.
Most of our everyday reasoning is defeasible. For example, when I
infer that the book on my desk is red because it looks red to me, I am
reasoning defeasibly. That is, I could be wrong without being wrong about
whether it looks red to me. When I infer that my wife is angry because she
is yelling at me and brandishing an umbrella, I could be wrong—she might
be pretending. When physicists proposed the Bohr atom as the best expla-
THE FORMAL STUDY OF REASONING 11

nation for various phenomena in atomic physics, they were wrong, as we


now know. That is, their inference was a reasonable one, but subsequent
discoveries have forced the retraction of their conclusion.
We do employ deductive reasoning, but it is usually interspersed with
defeasible reasoning. For example, consider the argument to the effect that
John is the murderer. That argument is deductive, but where do the premises
come from? The first premise was to the effect that the murderer was either
John or Joe, and was based upon eyewitness testimony. Clearly, such testi-
mony provides only a defeasible basis for an inference. The same is true of
the other premises of that argument. The picture of human reasoning that
emerges from this is one in which we engage in a lot of defeasible reasoning,
drawing conclusions on the basis of perceptual input and induction, and
then once we have a fair amount of information to go on, we do some
deductive reasoning.
This suggests that if we want to study reasoning, we will learn more
by studying defeasible reasoning rather than deductive reasoning. That is
probably true, but defeasible reasoning is harder to study than deductive
reasoning and less well understood. It is also easier to understand the way
defeasible reasoning works against a background understanding of deductive
reasoning. So the bulk of this book will be concerned with deductive rea-
soning, and only at the end will we return to an examination of defeasible
reasoning.

5. Deductive Inference
5.1 Logical Possibility
It is customary to try to define deductive inference in terms of logical
possibility. This is the way we proceeded when we introduced the term in
section four. We said that a deductive inference is one in which the conclusion
could not possibly be false while the premises are true.
There are many kinds of possibility. There is an “epistemic” sense in
it is possible that the President is now in Honduras but it is not possible
that he is now on Mars. This is epistemic in the sense that, “for all we
know”, he might be on Honduras, but our knowledge rules out the possibility
of his being on Mars.
There is also a non-epistemic sense in which, even when we know
that the President is currently in Washington, we are willing to say that the
President could now be in Honduras but could not now be on Mars. This
has to do with “practical possibility”. The President could have gone to
Honduras, but he could not have gone to Mars even if he wanted to.
Another variety of possibility is physical possibility. Something is
physically possible just in case it is not ruled out by fundamental laws of
nature. For example, the special theory of relativity tells us that nothing
can move faster than the velocity of light, so traveling faster than light is
physically impossible. On the other hand, traveling at 99.999% the velocity
THE FORMAL STUDY OF REASONING 12

of light is physically possible, but certainly not practically possible.


Physical possibility is a very weak requirement. Almost anything we
can think of is physically possible, even if it isn’t practically possible.
However, there is an even weaker kind of possibility. Traveling faster than
light is physically impossible, but the laws of nature might have been other
than they are, in which case something could travel faster than light. The
sense in which the laws of nature might have been other than they are is
logical possibility. Logical possibility is the weakest kind of possibility. All
that it precludes are those things that could not, in any sense, be true.
A popular way of trying to explain logical possibility is in terms of
possible worlds. A possible world is “a way the world could be”. A possible
world is specified by giving a complete description of a world, where that
description is logically possible. Then something is logically possible just in
case there is a possible world in which it is true. Of course, this explanation
is circular, because we need the notion of logical possibility to define the
notion of a possible world. Nevertheless, this is sometimes a helpful way of
thinking about logical possibility.

5.2 Logical Necessity and Validity


Logical necessity is defined in terms of logical possibility. Something is
logically necessary just in case it is not logically possible for it to be false.
Equivalently, something is logically necessary just in case it is true in all
possible worlds. Logical necessities are things that “couldn’t be false”, in
the strongest possible sense of “couldn’t”.
It is customary to characterize deductive inference in terms of logical
necessity or logical possibility. An inference is deductive just in case it is
logically impossible for the premises to be true while the conclusion is false.
Equivalently, an inference is deductive just in case it is logically necessary
that if the premises are true then the conclusion is true.
An argument encodes an inference or sequence of inferences. A deductive
argument is one in which all of the inferences are (or purport to be) deductive.
If all of the inferences are deductive, it follows that if the premises of the
argument are true then the conclusion is also true. This relationship between
the premises and conclusion are captured by the concept of a valid argument.
Consider an argument having some number, say n, premises: P1, P2 , P 3, … ,
Pn. Let the conclusion be, Q. Then the argument looks as follows:

P1
P2
P. 3
..
Pn

Therefore, Q.

Note that when we write the argument this way, we are ignoring the
intermediate inferences. That is, we are focusing on the premises and
THE FORMAL STUDY OF REASONING 13

conclusion and abstracting from how the conclusion was gotten from the
premises. We say that such an argument is valid if, and only if, it is logically
impossible for its premises to be true while the conclusion is false. In other words,
the argument is valid if, and only if, the inference it encodes is deductive.
Corresponding to the preceding argument we can construct the
statement:
If P1 and P2 and P3 and . . . and P n, then Q.

This statement says that if the premises of the argument are all true then the
conclusion is true. This statement is called the corresponding conditional of
the argument. A conditional is any statement of the form “If P then Q.” The
corresponding conditional of an argument is the conditional in which P
says that the premises of the argument are all true and Q is the conclusion
of the argument. For example, the corresponding conditional of the following
argument:

1. All Republicans are faithful to their spouses.


2. No one who is faithful to his or her spouse will ever get into political
trouble.

Therefore, no Republican will ever get into political trouble.

is

If all Republicans are faithful to their spouses, and no one who is faithful
to his or her spouse will ever get into political trouble, then no Republican
will ever get into political trouble.

To say that an argument is valid is to say that it is logically necessary that if


its premises are all true then its conclusion is true; that is, its corresponding
conditional is logically necessary. Thus, for example, to say that the above
argument is valid is just to say that the statement “If all Republicans are
faithful to their spouses, and no one who is faithful to his or her spouse will
ever get into political trouble, then no Republican will ever get into political
trouble” is logically necessary. Hence we can talk about the validity of an
argument by talking about the logical necessity of its corresponding
conditional:

An argument is valid if, and only if, its corresponding conditional is logically
necessary.

If an argument is valid and its premises are true, it follows that its
conclusion is true. It should be emphasized, however, that if the premises
of a valid argument are false, it does not follow that the conclusion is also
false. The following is an example of a valid argument with (presumably)
false premises and a true conclusion:
THE FORMAL STUDY OF REASONING 14

1. All Communists are bank managers.


2. All bank managers are followers of Karl Marx.

Therefore, all Communists are followers of Karl Marx.

Given a valid argument, the only time we can infer anything about whether
its conclusion is true is when all its premises are true. If one or more of its
premises are false, then its conclusion can be either true or false. The only
combination of truth and falsity that is prohibited in a valid argument is all
premises being true and the conclusion being false. Any other combination
is possible.

5.3 Validity and Inferential Correctness


In section two it was asserted that logic studies the inferential correctness
of arguments. That is not the same thing as validity, even if we focus
exclusively on deductive arguments. This is because validity pertains only
to the relationship between the premises and the conclusion of an argument,
and ignores any intermediate steps whereby the conclusion is inferred from
the premises. To illustrate, suppose we expand the above argument by
adding an erroneous intermediate step and inferring the conclusion
(incorrectly) from the intermediate step:

1. All Communists are bank managers.


2. All bank managers are followers of Karl Marx.

3. Therefore, all followers of Karl Marx are Communists. (from 1 and


2)

Therefore, all Communists are followers of Karl Marx. (from 3)

The premises and conclusion are unchanged from the previous argument,
so the new argument is still valid. However, (3) does not follow from (1)
and (2), and the conclusion does not follow from (3), so the argument is not
inferentially correct. The argument is only valid “by accident”, because
two errors of reasoning cancelled out.
The term “sound” is sometimes used to refer to arguments that are
not only valid, but also through and through inferentially correct.
Unfortunately, that conflicts with the more customary use of “sound” to
refer to arguments that also have true premises, so we will not adopt that
terminology here. We might say instead that an argument is deductively
correct if, and only if, every inference in it is deductive (and hence inferentially
correct). If an argument is deductively correct, it follows that it is valid, but
as we have seen, an argument can be valid without being deductively correct.

5.4 Explaining Deductive Inference Functionally


There is a school of thought, deriving from W. V. Quine, that is critical
of the concepts of logical necessity and logical possibility. In effect, Quine’s
THE FORMAL STUDY OF REASONING 15

objection was that we cannot make clear sense of these concepts by providing
philosophical analyses of them. That is a rather strange objection, because
it could be applied equally to most of our fundamental concepts. For example,
philosophers have been no more successful at providing philosophical
analyses of the concept of a physical object, or the concept of time. Does
that mean that such concepts should be viewed as suspect? Surely not.
But regardless of whether the Quinean arguments are any good, they
have convinced a number of philosophers. Without the concepts of logical
necessity and logical possibility, we cannot explain deductive inference as
above. This throws the foundations of logic into disarray. However, there
is another way of characterizing deductive inferences. A deductive inference
is one for which there are no (could be no) defeaters. This is to characterize
deductive inference “functionally”, by describing how it works and what
role it plays in human cognition. This approach takes defeasible inference
to be the norm, and then characterizes deductive inference to be inference
that is not subject to defeat.
This approach cries out for an explanation of what determines whether
there are or could be any defeaters. That is to be explained by explaining
where the structure of reasoning comes from. My own conviction is that it
is dictated by the cognitive architecture of the human mind, which in turn
is determined by our physiology. But this is a long story, and I will not
pursue the details here.1
This approach provides a way of resurrecting logical necessity from
its Quinean coffin. The resurrection proceeds by turning the standard story
on its head and explaining logical necessity in terms of deductive inference
rather than the other way around. Deductive inference is characterized
functionally, as sketched above, and necessary truths are just those that can
be established purely deductively, without using any defeasible steps in
our reasoning.

5.5 Defeasible Correctness


It was remarked above that most of our reasoning is defeasible rather
than deductive. Typically, our reasoning will combine some defeasible
inferences and some deductive inferences. Because defeasible inferences
are not deductive, it is logically possible for the premises to be true but the
conclusion false. This is why defeasible inferences can be defeated. For
instance, when the inductive inference to “All swans are white” was defeated
by finding black swans in Australia, the defeat resulted from discovering
that the logical possibility of the premises being true but the conclusion
false was more than a mere possibility—the conclusion actually was false
despite the premises being true.
Arguments containing defeasible inferences cannot be expected to be

1
Some of the details of this story are spelled out in John Pollock, Cognitive
Carpentry, MIT Press, 1995; and in John Pollock and Joseph Cruz, Contemporary Theories of
Knowledge, second edition, Rowman and Littlefield, 1999.
THE FORMAL STUDY OF REASONING 16

either valid or deductively correct. If they were they would be deductive


arguments rather than defeasible arguments. But this does not mean that
the arguments are somehow flawed. In the absence of defeaters, defeasible
arguments can make it reasonable to believe their conclusions. They just do
not logically guarantee that the conclusion is true given the premises. When
all of the steps of a defeasible argument are inferentially correct, let us say
that the argument itself is defeasibly correct. Defeasible correctness is to
defeasible arguments as deductive correctness is to deductive arguments.
The use of defeasible arguments involves a trade-off. Defeasible
arguments make their conclusions less certain than deductive arguments do.
A deductive argument absolutely guarantees the truth of the conclusion if
the premises are true. With a defeasible argument, it remains logically
possible for the conclusion to be false even if the premises are true. The
defeasible argument makes it reasonable to expect that the conclusion is
true, but does not guarantee its truth. In this respect, we might think of
defeasible arguments as inferior to deductive arguments.
But there is another sense in which they are superior, namely, they
give us more information. A deductive argument can never lead to a
conclusion that isn’t already implicit in the premises. In effect, what a
deductive argument does is analyze the premises and reveal the consequences
of what is being assumed. Defeasible arguments, on the other hand, are
ampliative, in the sense that they can lead us to genuinely new information.
For example, when we infer that no earthworms have wings, we can then
deduce that earthworms in China don’t have wings, even though our original
evidence for the conclusion that no earthworms have wings did not include
any earthworms in China. Similarly, when I infer that the book on my desk
is red because it looks red, the conclusion that the book is red expands
upon my evidence and tells me something new about the world. It can be
argued convincingly that the use of defeasible inference is essential to our
being able to acquire knowledge of the world.2 Without defeasible reasoning
we would never get beyond judgments about how things look to us.

6. Formal Logic
6.1 Argument Forms and Formal Validity
It was remarked above that an argument can be inferentially correct
either because of the particular concepts that are employed in it or because
of its general structure. When an argument is inferentially correct for the
latter reason, we say that it’s form makes it correct.
The form of an argument is obtained by replacing some of its terms by
schematic letters. For example, we might write a simple argument form as:

2
See Pollock and Cruz, Contemporary Theories of Knowledge.
THE FORMAL STUDY OF REASONING 17

1. All A are B.
2. No B is C.

Therefore, no A is C

Here A, B, and C are schematic letters whose places can be occupied by


arbitrary predicates. What is important about this argument form is that
every argument of this form will be inferentially correct, regardless of how
we instantiate A, B, and C.
We defined an argument to be valid if, and only if, it is logically necessary
that the conclusion is true provided the premises are true. We can extend
the concept of validity to argument forms, saying that an argument form is
valid if, and only if, every argument of that form is valid. Logic is often
called “formal logic” because it concerns itself with the validity of argument
forms rather than arguments. In other words, logic is concerned with the
structural features of an argument that can make it valid (and hence make
all other arguments with the same structure valid). The nonstructural aspects
of arguments that can make them either valid or inferentially correct are
also of philosophical interest, but they fall under the purview of epistemology
rather than logic.

6.2 Statement Forms and Formal Necessity


Just as the form of an argument can make it valid, the form of a
statement can make it logically necessary. For example

Either it is snowing or it is false that it is snowing

is logically necessary. However, the source of its necessity has nothing to


do with meteorology. Any statement of the form

Either P or it is false that P

will be logically necessary. The source of the necessity is the form of the
statement rather than the content of its constituents.
Statements do not have unique logical forms. Consider the statement

If John comes and Mary stays home then Bill be will unhappy.

This statement can be regarded as having at least two different forms:

If P and Q then R.
If S then R.

A statement might have one logical form all of whose instances are logically
necessary, and another (more general) logical form some instances of which
are not logically necessary. For example, consider the statement
THE FORMAL STUDY OF REASONING 18

If all featherless bipeds are men, and all men are mortal, then all
featherless bipeds are mortal.

This statement has the logical form:

If all A are B and all B are C then all A are C.

Every statement of this form is logically necessary. But it also has the more
general form:

If P then Q.

Statements of this form need not be logically necessary.


Let us say that a statement form is formally necessary if, and only if, every
statement of that form is logically necessary. If a statement has a formally
necessary form, then the statement is logically necessary. However, state-
ments can also be logically necessary because of relationships between their
constituents that are not captured by logical form. For instance, the statement

All bachelors are unmarried

is logically necessary, but it does not exhibit a form that is formally necessary.

6.3 Formal Necessity and Formal Validity


Logic studies both formal necessity and formal validity. There is an
important connection between these two concepts. The validity of arguments
(as opposed to argument forms) was defined in terms of the logical necessity
of the corresponding conditional. We can also talk about the corresponding
conditional of an argument form. This will be a statement form. For example,
the corresponding conditional of the argument form

1. All A are B.
2. No B is C.

Therefore, no A is C

is

If all A are B and no B is C then no A is C.

In relating argument forms to their corresponding conditionals, the


crucial observation is that a statement has the form of the corresponding
conditional of the argument form if, and only if, it is the corresponding
conditional of an argument of that form. For instance, the following argument
and its corresponding conditional instantiate the preceding argument form
and its corresponding conditional:
THE FORMAL STUDY OF REASONING 19

1. All A are B. 1. All ducks have feathers.


2. No B is C. 2. Nothing having feathers is a mammal.

Therefore, no A is C Therefore, no ducks are mammals.

If all A are B and no If all ducks have feathers and nothing having
B is C then no A is C. feathers is a mammal then no ducks are mammals.

So a statement has the form of the corresponding conditional of the


argument form if, and only if, it is the corresponding conditional of an
argument of that form. An argument form is valid if, and only if, every
argument of that form is valid. Any particular argument of that form is
valid if, and only if, its corresponding conditional is logically necessary. So
all arguments of a form are valid if, and only if, every statement instantiating
the corresponding conditional is logically necessary. But the latter means,
by definition, that the statement form that is the corresponding conditional
of the argument form is formally necessary. So in sum:

An argument form is valid iff its corresponding conditional is formally


necessary.

In light of this connection between argument forms and their cor-


responding conditionals, one way to study valid argument forms is to study
the formal necessity of their corresponding conditionals. Most logical theories
make use of this connection.

7. Logical Theories
7.1 The Implicational Calculus
Logic is the study of valid argument forms and formally necessary
statement forms. It was remarked above that there is no such thing as the
form of a statement or argument. Statements and arguments can exemplify
many different forms. Different logical theories are generated by focusing on
different classes of statement and argument forms. To take a slightly simplistic
example, consider all statement and argument forms that can be constructed
using just “and”, “if ... then”, punctuation, and schematic letters standing
for statements. Examples of statement forms constructed using these mate-
rials are:

If P and Q then R.
If, if P then Q and if Q then R, then if P then R.

As an illustration of how logic investigates logical forms, let us see what a


logical theory of these logical forms would look like. Let us call the theory
the implicational calculus. This is not a theory that people actually study, but
THE FORMAL STUDY OF REASONING 20

we can use it to illustrate how a logical theory proceeds.

Syntax
The first step in constructing a logical theory is generally to introduce
an artificial representation of the statement forms studied. We could express
them in English (augmented with schematic variables), but there are two
problems with that. First, they would be hard to read. Second, English
sentences are often ambiguous, and it is desirable to have a notation that
avoids that ambiguity. For example, consider the expression we wrote
above:

If, if P then Q and if Q then R, then if P then R.

This could mean several different things depending upon how the parts of
the expression are grouped together. We could disambiguate it using pa-
rentheses for punctuation, in either of the following ways:

If [(if P then Q) and (if Q then R)] then (if P then R).

If {if P then [Q and (if Q then R)]} then (if P then R).

The first reading is probably the more natural, but the rules of English do
not preclude the second reading. We want to avoid such ambiguities, and
the best way to do it is to use parentheses for punctuation, just as we did
above.
The English words “and” and “if ... then” also turn out to be ambiguous.
For example, “P and Q” can express simple conjunction, saying merely that P
and Q are both true, or it can express a temporal relation as in “He lay
down and fell asleep”. As a simple conjunction, “P and Q” is equivalent to
“Q and P”, but “He lay down and fell asleep” means something different
from “He fell asleep and lay down”. To avoid this source of ambiguity, it is
customary to introduce artificial symbols to replace the English words, and
rule that these symbols have some particular interpretation of the English
words as their meanings. The standard procedure is to symbolize “P and
Q” as (P & Q), taking “&” to express what I called “simple conjunction”.
“If P then Q” is symbolized as (P → Q). Using these devices, we can
symbolize the two readings of “If, if P then Q and if Q then R, then if P then
R” as:

[(P → Q) & (Q → R)] → (P → R)

{P → [Q & (Q → R)]} → (P → R).

The expressions we can write in this way using “&”, “→”, schematic letters,
and parentheses are called formulas of the implicational calculus.
In general, in constructing a logical theory, the first step is to define an
artificial symbolism and define formulas of the theory to be the expressions of
THE FORMAL STUDY OF REASONING 21

the symbolism that express statement forms studied by the theory. The
formulas will be constructed out of logical symbols, in this case “&” and
“→”, schematic letters, and parentheses. This part of the logical theory is
its syntax.

Semantics
The next step in constructing a logical theory is to define what the
formulas mean, i.e., to explain precisely what statement forms are expressed
by any given formula. This explanation constitutes the semantics of the
theory. The general procedure is to define the notion of an interpretation,
and then explain the meaning of a formula by saying which interpretations
make it true.
The definition of an interpretation varies a great deal from one logical
theory to another. In general, interpretations are defined so that they abstract
from the entire meaning and just specify the minimal amount of information
needed to determine whether a formula is true or false. In the implicational
calculus, it turns out that all we have to know to determine the truth value
of a formula (i.e., whether it is true or false) is the truth values of the
schematic letters out of which it is built. So we can define an interpretation
of the implicational calculus to be an assignment of truth values (truth or
falsity) to schematic letters. Other logical theories may have to include
more or different information in their interpretations.
Given a definition of the interpretations of the logical theory, formulas
receive their meanings by specifying the rules that determine under which
interpretations they are true. The truth rule for “&” is obvious:

(P & Q) is true if, and only if, P is true and Q is true.

The truth rule for “→” is not obvious. The following rule is normally used:

(P → Q) is true if, and only if, it is not the case that Q is false but P is
true.

This rule will be discussed at length in the next chapter. For now it will just
be used as an illustration of the kind of rule one might adopt.
Repeated application of these rules enables us to determine the truth
value of any formula of the implicational calculus under any interpretation.
For example, suppose P is assigned truth, and Q and R are assigned falsity.
Then consider the formula

[(P → Q) & (Q → R)] → (P → R).

The truth rule for “→” makes (P → Q) and (P → R) false, but makes (Q →
R) true. The rule for “&” then makes [(P → Q) & (Q → R)] false. Finally,
the rule for “→” makes the whole formula true.
Different logical theories use different techniques for formulating the
semantics. In general, the desideratum is to describe the semantics in a
THE FORMAL STUDY OF REASONING 22

mathematically precise manner so that we can prove mathematical theorems


about it. This need not be very difficult, as the example of the implicational
calculus illustrates.
Formulas express statement forms. Once meanings have been assigned
to formulas, we can ask which formulas are formally necessary. If we have
done a good job giving a mathematical characterization of the meanings, it
will be possible to prove a theorem that tells us which formulas are formally
necessary. In fact, if we have defined interpretations appropriately, it will
almost always be the case that we have the following theorem:

A formula is formally necessary if, and only if, it is true under every
interpretation.

This theorem holds for the implicational calculus, although we will not
prove it here.
Formulas true under every interpretation are said to be valid. This is
unfortunate terminology, because it has no direct connection to the use of
“valid” in talking about arguments. However, it is the standard terminology,
so we will conform to convention and use it here. The reader is warned,
however, not to confuse the two meanings of “valid”.
Defining validity for formulas in this way, the preceding theorem can
be re-expressed as:

A formula is formally necessary if, and only if, it is valid.

Arguments
Our central interest in logic is usually with arguments rather than
statements. The third step in constructing a logical theory is to specify
what arguments can be constructed out of the formulas of the theory. Different
theories can deal with arguments having different structures. The simplest
arguments will be finite lists of formulas, constructed according to specified
rules of inference. Rules of inference are rules telling us that certain formulas
can be inferred from others. For example, in the implicational calculus we
might use the following rule of inference in constructing arguments:

From P and (P → Q), infer Q.

This simple rule of inference has a long history, and accordingly has a Latin
name, modus ponens (meaning “method of affirmation”).
Another simple rule that we might adopt is sometimes called adjunction:

From Q and R, infer (Q & R).

Let us also adopt a pair of rules that are collectively called simplification:

From (Q & R), infer Q.


From (Q & R), infer R.
THE FORMAL STUDY OF REASONING 23

Any bit of reasoning has to start with some premises, so we need a


rule telling us that we can introduce formulas as premises. We normally
want to allow the use of arbitrary additional premises in arguments so that
we can study argument forms proceeding from those premises. Then we
define a derivation in the implicational calculus to be a finite list of formulas
each formula of which is either a premise or can be inferred from previous
formulas in the list using the rules of inference we have chosen. Thus, for
example, the following would be a derivation in the implicational calculus:

1. (P & Q) (premise)
2. (P → R) (premise)
3. (Q → S) (premise)
4. P (simplification, from 1)
5. Q (simplification from 1)
6. R (modus ponens, from 2 and 4)
7. S (modus ponens, from 3 and 5)
8. (R & S) (adjunction, from 6 and 7)

Derivations are argument forms. They are the argument forms studied
by the logical theory in which they occur. A derivation is said to be a
derivation of whatever formula appears on its last line from whatever premises
are used in it. Thus the preceding derivation is a derivation of (R & S) from
(P & Q), (P → R), and (Q → S).
We have seen that we can evaluate argument forms in terms of their
corresponding conditionals. An argument form is valid if, and only if, its
corresponding conditional is formally necessary. If the formal necessity of
formulas corresponds to validity, then an argument form is valid if, and
only if, its corresponding conditional is a valid formula. (Note the two
distinct senses of “valid” in the preceding sentence.) Hence the semantics
we have adopted for a logical theory provides the basis for investigating
the validity of argument forms constructed within the theory.
To facilitate the investigation of argument forms within a logical theory,
we define:

A set of formulas P1 ,...,Pn entails a formula Q if, and only if, the
corresponding conditional (P1 &...&Pn) → Q is valid.

An argument is evaluated by asking whether its premises entail its conclusion.


We typically want our derivations to satisfy two conditions:

Soundness: If a formula is derivable from a set of premises, the premises


should entail the formula.

Completeness: If a set of premises entails a formula, the formula


should be derivable from the premises.

The rules given above for the implicational calculus are sound but not
complete. Notice that the use of “sound” in talking about systems of der-
THE FORMAL STUDY OF REASONING 24

ivations is unrelated to its use in talking about argument (i.e., requiring


arguments to have true premises).

7.2 The Propositional and Predicate Calculi


The implicational calculus is too simple to be of much real interest.
Our first serious excursion into logic will consist of an investigation of the
propositional calculus. The propositional calculus is much like the implicational
calculus except that it adds logical symbols for “it is false that”, “or”, and
“if and only if”. We can express some fairly complex reasoning within the
propositional calculus, and the propositional calculus is still sufficiently
simple to constitute a good introduction to logic. However, ultimately, it is
not very useful. Most of the deductive reasoning we actually perform
requires a richer symbolism.
After discussing the propositional calculus, we will turn to the predicate
calculus. The predicate calculus is constructed by adding additional logical
symbols to the propositional calculus. Specifically, we add ways of expressing
“all” and “some”. It turns out that almost all of our deductive reasoning
can be expressed in the predicate calculus. It is a very powerful theory.
However, its semantics and systems of derivations are correspondingly more
complex than for the propositional calculus. Fortunately, in studying the
predicate calculus, we can build smoothly on what we learn from the prop-
ositional calculus. Both theories follow the general line illustrated by the
implicational calculus, but the notion of an interpretation is more complex
for the predicate calculus and the rules of inference are also more complex.

8. Summary

We defined the following terms:

Reasoning is inferentially correct if, and only if, given the premises, the
reasoning supports the conclusion. This is independent of whether the
premises are actually true.

A sound argument is one that is inferentially correct and also has true
premises.

An inference is deductive if, and only if, it is logically impossible for the
conclusion to be false when the premises are true.

Defeasible reasoning is reasoning that is inferentially correct but not


deductive.

An argument is valid if, and only if, it is logically impossible for its
conclusion to be false when its premises are true, i.e., if, and only if, the
inference it encodes is deductive.
THE FORMAL STUDY OF REASONING 25

The corresponding conditional of an argument is the statement that if the


premises are true then the conclusion is true.

It was then shown that we can evaluate the validity of an argument by


looking at its corresponding conditional:

An argument is valid if, and only if, its corresponding conditional is


logically necessary.

Turning to argument forms and statement forms, we defined:

An argument form is valid if, and only if, every argument of that form is
valid.

A statement form is formally necessary if, and only if, every statement of
that form is logically necessary.

It was then shown that we can evaluate the validity of an argument form by
looking at its corresponding conditional:

An argument form is valid iff its corresponding conditional is formally


necessary.

A logical theory focuses on a particular class of statement forms and investigates


the formal necessity of statement forms in that class and the validity of
argument forms constructed out of the statement forms.

A logical theory usually has three parts:

The syntax introduces an artificial symbolism for expressing the statement


forms unambiguously, and defines what a formula is.

The semantics explains the meaning of the formulas. It does this by defining
what an interpretation is and giving truth rules that determine which
formulas are true relative to an interpretation. In the implicational calculus,
an interpretation is an assignment of truth values to the schematic letters.

We defined:

A formula is valid if, and only if, it is true under every interpretation.

A set of formulas P1,...,P n entails a formula Q if, and only if, the
corresponding conditional (P1 &...&Pn) → Q is valid.

If we have defined interpretations appropriately, it will almost always


be the case that we have the following theorem:
THE FORMAL STUDY OF REASONING 26

A formula is formally necessary if, and only if, it is true under


every interpretation.

The system of derivations consists of a precise set of rules for constructing


arguments using formulas. The arguments constructed in accordance
with the rules are derivations. We typically want our derivations to satisfy
two conditions:

Soundness: If a formula is derivable from a set of premises, the premises


should entail the formula.

Completeness: If a set of premises entails a formula, the formula


should be derivable from the premises.
Part One:

The
Propositional
Calculus
2
The Syntax of the
Propositional Calculus

1. Introduction to the Propositional Calculus


The propositional calculus is an extension of the implicational calculus.
The implicational calculus studied those statement and argument forms
that can be constructed using just “and”, “if ... then”, punctuation, and
schematic variables standing for statements. The formulas of the implicational
calculus are not sufficiently expressive to be of much interest. However, if
we augment them slightly we get a logical theory of much more interest.
For this purpose we need to add “or”, “if and only if”, and “it is false that”
to the list of constituents of statement forms. The resulting theory, the
propositional calculus, is both of historical interest and of some practical interest.
It is of historical interest because it is the first modern logical theory to have
been studied carefully, and it is of practical interest because within the
propositional calculus we can represent a number inference patterns that
we employ regularly in our everyday reasoning.

2. The Sentential Connectives


The expressions “and”, “or”, “if … then”, “if and only if”, and “it is
false that”, are called sentential connectives, because they are used to connect
sentences to form larger sentences. For example, the statement

If it is false that it is either raining in the mountains or snowing in


the mountains then it is false that it is raining in Tucson

has the form

If it is false that (R or S) then it is false that T

where now R stands for “It is raining in the mountains”, S stands for “It is
snowing in the mountains”, and T stands for “It is raining in Tucson”.

2.1 Logical Symbols


It was observed in chapter one that English punctuation is often
ambiguous, and we introduced the use of parentheses to avoid the ambiguity.
We follow that course again in the propositional calculus. It was also
observed that the actual English words that we have called “the sentential
connectives” are ambiguous, having several different usages. To avoid this
ambiguity, we will follow the same strategy used in the implicational calculus.
That is, we will introduce artificial symbols to express just that meaning of
each sentential connective that we want to study in the propositional calculus.
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 30

The symbols we use to replace the sentential connectives are called the
logical symbols of the propositional calculus. They are as follows:

Symbol Meaning

~P it is false that P
(P & Q) P and Q (or, both P and Q)
(P ∨ Q) either P or Q
(P → Q) if P then Q
(P ↔ Q) P if and only if Q

We will have to elaborate on the explanations of meaning given in this table


in order to avoid the ambiguities that are built into English, but let us
postpone that for the moment.
Statements formed using the logical symbols have names, as follows:

• ~P is the negation of P.
• (P & Q) is the conjunction of P and Q, and P and Q are the conjuncts.
• (P ∨ Q) is the disjunction of P and Q, and P and Q are the disjuncts.
• (P → Q) is a conditional, P is the antecedent, and Q is the consequent.
• (P ↔ Q) is a biconditional. There are no standard names for the left
and right side of a biconditional.

2.2 Sentential Letters


Using these symbols we can represent the forms of quite complex
statements. We have been using capital letters as the schematic variables
symbolizing the simple statements that form the parts of compound
statements. We might encounter a statement having more than 26 sentential
parts, so to accommodate this theoretical possibility we can also use capital
letters with subscripts, such as P7 or Q13. In this book we will not actually
encounter any sentences that long, but it may still be useful to have schematic
variables with subscripts. When symbolizing the form of a statement, it is
often helpful to use letters that remind us of what they are being used to
symbolize. For example, we might symbolize “It is raining” as R, and “It is
snowing” as S. However, the use of such mnemonics will occasionally
leave us wanting to use the same letter to symbolize two different sentences.
We can get around that by using subscripts.
It should be emphasized that the same letter with different subscripts,
such as P and P7, can be used to represent totally unrelated statements. The
fact that P occurs in P7 is just a coincidence. Capital letters, with or without
subscripts, used this way as schematic variables are called sentential letters.
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 31

3. Formulas of the Propositional Calculus


Using sentential letters to stand for statements, and then combining
them with logical symbols, we can express quite complex statement forms.
The expressions we construct in this way are called formulas of the propositional
calculus. We can give precise rules for constructing formulas of the
propositional calculus. The simplest formulas are simply sentential letters,
with or without subscripts, such as P, Q, P13, R129, and so forth. We call
these atomic formulas because they are the atoms from which more complicated
formulas are constructed. Then all other formulas of the propositional
calculus can be constructed by successive applications of the following rules:

1. An atomic formula is a formula.


2. If A is any formula, then ~A is a formula.
3. If A and B are any formulas, then (A & B) is a formula.
4. If A and B are any formulas, then (A ∨ B) is a formula.
5. If A and B are any formulas, then (A → B) is a formula.
6. If A and B are any formulas, then (A ↔ B) is a formula.

In writing the rules in this way, A and B are used as variables to stand for
arbitrary formulas of the propositional calculus. Such variables are called
metalinguistic variables.
These rules define the concept of a formula of the propositional calculus.
That is, an expression is a formula of the propositional calculus if, and only
if, it can be constructed by repeated application of these rules. To illustrate,
consider the formula

(P → (~Q & R))

We can construct this formula using the above six rules, as follows. We
begin with the smallest parts and work outwards. By Rule 1, P, Q, and R
are formulas. Then by Rule 2, ~Q is a formula. By Rule 3, as ~Q and R are
both formulas, (~Q & R) is a formula. Then by Rule 5, as P and (~Q & R)
are both formulas, (P → (~Q & R)) is a formula.
We can construct very complicated formulas using these rules. For
example, consider the formula
((P → (Q3 ↔ ~R4 )) ↔ ~~(~P & ~( Q3 ↔ (R4 ∨ (R17 & ~~P)))))
Let us see how we would construct this formula using the six rules. Again,
we begin with the smallest parts and work outwards. By Rule 1:
Q3, R4 , R17
are formulas because they are atomic formulas. As P and R4 are formulas,
we can use Rule 2 to construct the formulas:
~P, ~R4 .
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 32

As ~P is then a formula, by Rule 2 again,


~~P
is a formula. Then as R17 and ~~P are both formulas, by Rule 3,
(R17 & ~~P)
is a formula. Then as R4 and (R 17 & ~~P) are both formulas, by Rule 4,
(R4 ∨ (R17 & ~~P)) is a formula.
Then as Q3 and (R4 ∨ (R17 & ~~P)) are both formulas, by Rule 6,
(Q3 ↔ (R4 ∨ (R17 & ~~P)))
is a formula. Then by Rule 2,
~(Q3 ↔ (R4 ∨ (R17 & ~~P)))
is a formula. We have already seen that ~P is a formula, so by Rule 3,
(~P & ~(Q3 ↔ (R4 ∨ (R17 & ~~P))))
is a formula. Then by Rule 2,
~(~P & ~(Q3 ↔ (R4 ∨ (R17 & ~~P))))
is a formula, and then by Rule 2 again,
~~(~P & ~(Q3 ↔ (R4 ∨ (R17 & ~~P))))
is a formula. This gives us the right side of the biconditional. Now working
on the other side, as Q3 and ~R 4 are both formulas, by Rule 6,
(Q3 ↔ ~R4 )
is a formula. P is a formula, so by Rule 5,
(P → (Q3 ↔ ~R4 ))
is a formula. Both sides of the biconditional have now been constructed,
and then by Rule 6,
((P → (Q3 ↔ ~R4 )) ↔ ~~(~P & ~(Q3 ↔ (R4 ∨ (R17 & ~~P)))))
is a formula. We can diagram the construction of this formula as follows:
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 33

P Q3 R4 P Q3 R4 R17 P Rule 1
~R4 ~P ~P Rule 2
~~P Rule 2
(R17 & ~~P)) Rule 3
(R4 ∨ (R17 & ~~P)) Rule 4
(Q3 ↔ (R4 ∨ (R17 & ~~P))) Rule 6
~(Q3 ↔ (R4 ∨ (R17 & ~~P)))) Rule 2
(~P& ~(Q3 ↔ (R4 ∨ (R17 & ~~P)))) Rule 3
~(~P& ~(Q3 ↔ (R4 ∨ (R17 & ~~P)))) Rule 2
~~(~P& ~(Q3 ↔ (R4 ∨ (R17 & ~~P)))) Rule 2
(Q3 ↔ ~R4 ) Rule 6
(P → (Q3 ↔ ~R4 )) Rule 5
((P → (Q3 ↔ ~R4 )) ↔ ~~(~P& ~(Q3 ↔ (R4 ∨ (R17 & ~~P))))) Rule 6

Notice that whenever we construct a conjunction, disjunction,


conditional, or biconditional, we enclose it in parentheses. In order to make
formulas easier to read, parentheses are often replaced by brackets and
braces. Thus we might write the above formula as
{[P → (Q3 ↔ ~R4 )] ↔ ~~[~P & ~(Q3 ↔ {R4 ∨ (R17 & ~~P)})]}
By doing this we can tell at a glance which left parenthesis or bracket goes
with which right parenthesis or bracket; thus, we can see more easily what
the formula means. But it is to be emphasized that the expression we
obtain by using brackets and braces is not literally a formula—it is an
abbreviation for a formula.
Formulas with large numbers of nested parentheses can be hard to
read. A useful trick for parsing such formulas is to connect matching
parentheses with lines:

((P → (Q3 ↔ ~R4 )) ↔ ~~(~P& ~(Q3 ↔ (R4 ∨ (R17 & ~~P)))))

Atomic formulas are the simplest formulas. Let us say that a formula
is molecular if it is not atomic. Then molecular formulas are those formulas
that contain logical symbols.
Sometimes we will want to talk about one formula being a part of
another formula. This just means that the first occurs somewhere in the
second. For example, P, Q, R, (P → Q), ~(P → Q), and [~P → Q) ∨ R] are all
parts of the formula ~[~(P → Q) ∨ R]. Let us also adopt the convention of
saying that a formula is part of itself. Then

~[~(P → Q) ∨ R]

is also a part of ~[~(P → Q) ∨ R]. The atomic parts of a formula are those
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 34

parts of the formula that are atomic. Thus P, Q, and R are the atomic parts
of ~[~(P → Q) ∨ R], while (P → Q), ~(P → Q), [~(P → Q) ∨ R] and ~[~(P →
Q) ∨ R] are parts that are not atomic.
With the exception of negations and atomic formulas, all formulas
must have outer parentheses. However, the outer parentheses are not
necessary for making the formula unambiguous. The outer parentheses are
only needed when the formula is used in constructing a larger formula. For
example,
P→Q
is unambiguous. However, to form its negation we cannot write
~P → Q.
That would be the abbreviation for (~P → Q), whereas what we want is ~(P
→ Q). So for the sake of enhanced readability, we could allow ourselves to
abbreviate formulas by omitting outer parentheses. But if we do that we
must take care to supply the omitted parentheses when combining formulas
with other formulas. And it must be borne in mind that this is strictly an
informal abbreviation, and what we write in that way is technically not a
formula.

Exercises

A. Show how the following formula of the propositional calculus can be


constructed using Rules 1 through 6. Diagram the construction as was
done on page 30.

((P → Q) ↔ (~(Q → P) ∨ ~~(Q & ~P)))

B. Construct a table for each of the following expressions indicating: (1)


whether it is a formula of the propositional calculus; (2) if it is a formula,
whether it is atomic or molecular; (3) if it is a molecular formula, whether it
is a negation, conjunction, disjunction, conditional or biconditional; (4) if it
is a formula, what its atomic parts are; (5) if it is a formula, what its parts
are:
1. P
2. ~~P
3. P∨Q
4. ((P ∨ Q) → ~Q)
5. (~(P ↔ Q))
6. (P ∨ Q → R)
7. ~((Q & R) ↔ (~Q ∨ ~R))
8. (P ∨ P)
9. (P & ~P)
10. (P → Q) ↔ ~(Q → P)
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 35

4. Symbolizing Statement Forms


Formulas of the propositional calculus can be used to symbolize quite
complex statement forms. For example, consider the statement
If he comes we will have the party at his house, and if he doesn't come
then we will have the party at Jones’ house.
Letting P be the statement “He comes”, Q the statement “We will have the
party at his house”, and R the statement “We will have the party at Jones’
house”, we can first symbolize the statement as
If P then Q, and if it is false that P then R.
This in turn can be symbolized as
If P then Q, and if ~P then R
and then as
(P → Q) and (~P → R)
and finally as
[(P → Q) & (~P → R)].
As another example, consider
If Jones needs money, then either he will reduce prices or he will apply
to the bank for a loan.
Letting P be “Jones needs money”, Q be “He will reduce prices”, and R be
“He will apply to the bank for a loan”, this statement can be symbolized,
first as
If P then either Q or R,
then as
If P then (Q ∨ R),
and finally as
[P → (Q ∨ R)].
There are two things that should be noticed about the preceding
examples. The first is that in symbolizing relatively complex statements
like “If he comes we will have the party at his house, and if he doesn't come
then we will have the party at Jones’ house”, we begin by symbolizing the
smallest parts, and then we construct the successively larger parts one at a
time until we finally get the whole statement. The symbolization went as
follows:
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 36

If P then Q , and if it is false that P then R


(P → Q) and if ~P then R
(P → Q) and if ~P then R
(P → Q) and (~P → R)
(P → Q) and (~P → R)
[(P → Q) & (~P → R)]

It is always best to begin by symbolizing the smallest parts of a statement


first, and then constructing the symbolization of the successively larger
parts one at a time in terms of the symbolization of their constituents. It is
unwise to try to symbolize an entire statement in one step if the statement is
at all complicated.
The second thing to notice is the use of parentheses. Whenever we
symbolize a part of the statement—unless that part is a negation—we enclose
it in parentheses to keep it separate from the other parts of the statement.
The parentheses take the place of commas and other grammatical devices of
English. We must be careful to include the parentheses, or the resulting
symbolization will be ambiguous. Consider the two statements
It is not true that Jones has a girlfriend and his wife is going to divorce
him
and
It is not true that Jones has a girlfriend, and his wife is going to divorce
him.
These obviously mean quite different things. The first denies that it is true
both that Jones has a girlfriend and that his wife is going to divorce him,
whereas the second says that Jones does not have a girlfriend, but his wife
is going to divorce him anyway. These two statements would be symbolized
as ~(P & Q) and (~P & Q) respectively. But if we omitted the parentheses in
symbolizing these statements and just wrote ~P & Q, it would be
indeterminate which of these two different statements were meant.
It is not necessary to enclose a negation (a statement of the form ~P) in
parentheses, because “ ~” does not really connect sentences—it acts on a
single sentence. But for any of the other sentential connectives it is necessary
to use parentheses each time the connective is used.

Exercises

Symbolize the forms of the following statements, paying careful attention


to the use of commas to force (sometimes unnatural) interpretations. Use
the sentential letters indicated:
1. If it rains today, the ground will be wet and we will not be able to
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 37

have a picnic. (R G P)
2. It is not true that if it is cloudy then it will rain. (C R)
3. If the sun shines in the morning it will rain, and if the sun does not
shine in the morning then it will not rain. (S R)
4. If we get plenty of sunshine, then if it rains the flowers will grow. (S
R F)
5. It is false that, there is a woman in the next room if and only if Jim
said there is. (W S)
6. It is false that there is a woman in the next room, if and only if, Jim
said there is. (W S)
7. Either the entrails will contain cockroaches or they will not, and if
they do then the gods are angered, and if they do not then Venus
will be in apposition to Jupiter. (E G V)
8. If she is an acrobat or she is a clown, then she lives in that trailer. (A
C T)
9. Harry will go to the bank today if and only if the market drops, and
if Harry goes to the bank today the supervisors will hide, and if the
supervisors hide and Harry does not go to the bank today then Emmett
will lose his shirt on the stock market. (H M S E)
10. If Francis Bacon wrote Hamlet and Shakespeare wrote Macbeth, then
either Shakespeare was Bacon or the theater manager was a crook.
(H M S C)

5. Conditionals
The sentential connective that has the greatest number of importantly
different uses is “if … then”. On any of its uses, it is said to express a
conditional, but there are several different kinds of conditionals.

5.1 Indicative Conditionals and Subjunctive Conditionals


The most important distinction is between indicative conditionals and
subjunctive conditionals. Indicative conditionals have the form “If it is true
that P then it is true that Q”. By contrast, subjunctive conditionals have the
form “If it were true that P then it would be true that Q”. These have
importantly different meanings. Ernest Adams1 illustrates the difference
with the pair of sentence:

If Oswald didn’t shoot Kennedy, someone else did.

If Oswald hadn’t shot Kennedy, someone else would have.

1
“The logic of conditionals”, Inquiry 8 (1965), 166-197.
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 38

The former is clearly true. After all, Kennedy got shot. But there is no
reason to think that the latter is true. Had Oswald been thwarted, Kennedy
would most likely have survived.
Most uses of subjunctive conditionals are to express what are called
counterfactual conditionals. A counterfactual conditional is about what would
have happened had something that is actually false been true. By contrast,
indicative conditionals are normally used to talk about what is true if
something we are not sure about is true.
There is an extensive literature on subjunctive conditionals, stemming
largely from Robert Stalnaker2 and David Lewis 3. My own account of
subjunctive conditionals can be found in my book The Foundations of
Philosophical Semantics4. There is no consensus on the analysis of subjunctive
conditionals, but it is agreed by everyone that however they are to be analyzed,
they are not properly expressed by the “→” of the propositional calculus.
Somewhat surprisingly, indicative conditionals have proved more
recalcitrant to analysis than subjunctive conditionals, and there is little
agreement about whether the “→” of the propositional calculus properly
expresses indicative conditionals.

5.2 Material Conditionals


In the propositional calculus, “→” expresses what is called the material
conditional. The material conditional (P → Q) is defined to mean (~P ∨ Q). In
the next chapter, we will see that this is equivalent to ~(P & ~Q), so the
latter could equally be used as the definition of the material conditional.
For example, if the sentence “If it rains then the crops will grow” is interpreted
as expressing a material conditional, it means “Either it won’t rain or the
crops will grow”. Equivalently, it means “It is false that it will rain but the
crops won’t grow”.
There is an argument which seems to show that the material conditional
is the same as the indicative conditional. Consider an indicative conditional
“If P then Q”. The argument purports to show that it is true if, and only if,
(~P ∨ Q) is true. For the sake of concreteness, let P be “It rains” and Q be
“The crops will grow”. The argument proceeds in two stages. First we
suppose that “If P then Q” is true and show that it follows that (~P ∨ Q) is
true. Then we suppose that (~P ∨ Q) is true and show that it follows that
“If P then Q” is true:
1. Suppose that “If P then Q” is true. So if it rains (i.e., P is true) then Q is
true. If it does not rain, then ~P is true. But either it will not rain or it will
rain. So either it will not rain, and then ~P will be true, or else it will rain,
and then Q will be true. Thus, no matter what happens, either ~P will be

2
“A theory of conditionals”, in N. Rescher, Studies in Logical Theory, Blackwell:
Oxford, 1968.
3
Counterfactuals, Harvard University Press: Cambrdige Mass, 1973.
4
Princeton University Press, 1984.
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 39

true or Q will be true; that is, (~P ∨ Q) will be true. So we see that if the
statement “If P then Q” is true, then (~P ∨ Q) is true.
2. In order to show that “If P then Q” and (~P ∨ Q) are true under exactly
the same circumstances, it must also be shown that if (~P ∨ Q) is true then
“If P then Q” is true. So let us suppose that (~P ∨ Q) is true. A disjunction
(A ∨ B) is true if, and only if, at least one of its disjuncts is true; that is, if,
and only if, either A is true or B is true. Thus if (~P ∨ Q) is true, then either
~P is true or Q is true. To show that “If P then Q” is true, we can reason as
follows:
If P is true, then ~P cannot be true. But by hypothesis, either ~P or Q
must be true. So if P is true then Q must be true; that is, “If P then Q” is
true.
Thus it has been shown that if (~P ∨ Q) is true, then “If P then Q” is true.
So we have an argument that seems to show that “If P then Q” and
(~P ∨ Q) are true in exactly the same circumstances. In other words, the
indicative conditional is properly analyzed as being the same as the material
conditional. However, as persuasive as the argument seems to be, the
conclusion is somewhat counterintuitive. The difficulty is that for a
disjunction to be true, all that is required is that at least one disjunct is true.
For instance, the disjunction “That is either gold or fool’s gold” is true if it is
gold and true if it is fool’s gold. This has the consequence that (~P ∨ Q) is
true if ~P is true, i.e., if P is false. But that is equivalent to saying that the
material conditional (P → Q) is true whenever its antecedent P is false. To
test this claim, consider an example. Suppose I have some influence with
the state Lottery Commissioner, but not enough to affect the outcome of the
lottery. I am also scrupulously honest and know that I would never use my
influence for any purpose. So I know that I will not use my influence with
the Lottery Commissioner. But from that I can infer that the material
conditional “If I use my influence with the Lottery Commissioner, you will
win the lottery” is true. If we understand this as an ordinary indicative
conditional, there is some temptation to deny that it is true, because my
influence is not so great that I could actually influence the outcome of the
lottery.
Examples like this are often taken to show that the material conditional
is not the same thing as the indicative conditional. However, we must not
be too quick to draw this conclusion. First, note that the most natural
interpretation of “If I use my influence with the Lottery Commissioner, you
will win the lottery” is as a subjunctive conditional, not an indicative
conditional. That is, what would ordinarily be taken as being asserted is
that if I were to use my influence, you would win the lottery. The distinction
between subjunctive and indicative conditionals is in terms of meaning, not
in terms of the words used. The words are ambiguous and can be used to
express either an indicative conditional or a subjunctive conditional.
Still, one can employ the sentence “If I use my influence with the
Lottery Commissioner, you will win the lottery” to express an indicative
conditional, and under the circumstances described there is something odd
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 40

about such a conditional. It does not seem reasonable to assert this just
because I know that I will not use my influence.
However, its being unreasonable to assert the conditional is not the
same thing as its being false. Under the circumstances described, it would
be equally odd to assert the disjunction “Either I won’t use my influence or
you will win the lottery”. The explanation for this may lie in what are
called “implicatures”.

5.3 Conversational Implicatures


The concept of an implicature is due to Paul Grice5. Logicians talk
about one statement implying another. That just means that it is logically
necessary that if the first statement is true then the second statement is true.
So understood, implication is a relation between statements. We also talk
about people implying things. One way to imply something is to make a
statement that implies it. But sometimes we can imply something by making
a statement that does not itself have that implication. Instead of the statement
having the implication, it is the act of making the statement under those
circumstances that has the implication. If my hostess asks me how I like the
dessert and I reply, “It is very pretty”, I imply that I don’t like the taste.
The statement that it is pretty does not imply that I don’t like the taste. My
making the statement that it is pretty, under those circumstances, is what carries
the implication. Similarly, if a professor critiques a student’s paper by
saying, “It is very well typed”, he implies that he does not like its content.
In cases like this, in which it is the act of making the statement rather than
the statement itself that carries the implication, we say that it is a conversational
implicature.
Grice observed that there are rules governing the correct use of language
that do more than determine meanings. Some of these rules govern
conversation. One such rule is something like, “Don’t say less than you
know that is relevant”. For example, if you are shown a metal bar and
asked what it is made of, if you know that it is made of titanium then there
would be something wrong with your replying, “It is made of either titanium
or aluminum”. By saying the latter you imply that you do not know which
it is made of. This is a conversational implicature.
One of the interesting characteristics of conversational implicatures is
that they can be “cancelled” by explicitly denying them. For instance, if I
say, “It is made of either titanium or aluminum, and I know which but I am
not going to tell you”, then I no longer imply that I don’t know which, nor
do I contradict myself. On the other hand, logical implications cannot be
cancelled. If I say, “There are more than four apples in that barrel, but I
don’t mean to imply that there are more than three”, I am simply talking
nonsense or contradicting myself.
Now let us apply this to material conditionals. It seems that what is
wrong with my asserting, “Either I won’t use my influence or you will win
the lottery”, is that I know that I won’t use my influence, and so I am saying
less than I know that is relevant. This is analogous to saying, “It is made of

5
“Logic and conversation”, in P. Cole and J. L. Morgan (eds.), Syntax and
Semantics, Academic Press: New York, 1975.
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 41

either titanium or aluminum”, when I know that it is made of titanium. I


am implying that I don’t know that I won’t use my influence, and so implying
that the disjunction is true because of some sort of connection between the
two disjuncts that makes one true if the other is false. Similarly, if I say, “If
I use my influence then you will win the lottery”, I am implying that I
might use my influence, and that the conditional is true because of some
connection between the antecedent and consequent such that if I did use my
influence then the consequent would be true.
Perhaps what is wrong with asserting an indicative conditional just
because the antecedent is false is not that the conditional isn’t true but
rather than it is a pointless thing to say. The point of asserting a conditional
is to license an inference. If we discover that the antecedent is true then we
can infer that the consequent is also true. Conversely, if we discover that
the consequent isn’t true, we can infer that the antecedent isn’t true. If we
already know that the antecedent isn’t true, then the first inference can
never occur, and the second inference is pointless because we already know
that its conclusion is true. Hence it is pointless to assert the conditional.
Grice’s observations suggest that there is nothing wrong with
identifying the indicative conditional with the material conditional. There
are cases in which the material conditional is true but it would be unreasonable
to assert the indicative conditional. However, Grice’s observations indicate
that it would be equally unreasonable to assert that material conditional.
Truth alone is not sufficient to make something assertable. The rules of
conversation impose other requirements as well, and the problematic
conditionals fail to satisfy those other requirements. That, rather than falsity,
seems to be their failing.

5.4 Using Material Conditionals


As we have seen, there are arguments to the effect that the indicative
conditional is the same as the material conditional. We should not regard
the matter as settled, however. This is still an issue being debated in
philosophical logic. Even if it were eventually decided, to everyone’s
satisfaction, that these two conditionals are not the same, we would continue
to use “→” to express the material conditional in the propositional calculus.
This is for two reasons. First, it is at least often true that we can express
what we want to express using material conditionals. Of course we cannot
express every conditional claim using material conditionals. For example,
we cannot express subjunctive conditionals in that way. But this just means
that the propositional calculus is not expressively complete, not that what it
can express is useless. Second, the material conditional turns out to be easy
to study. As we will see in the next chapter, it has a simple truth rule that
makes it possible to investigate its logical properties using the same semantics
that is required for “and”, “or”, and “it is false that”. These different logical
concepts form a nicely unified package that can be studied together, and
the result is the propositional calculus.
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 42

6. Paraphrasing
Sometimes it is necessary to paraphrase a statement before it can be
symbolized. Consider the statement
John and Joe both came to the party.
We want to symbolize this as a conjunction, but it cannot be symbolized
directly because “and” stands between two names rather than between two
sentences. Before we can symbolize it we must paraphrase it so that the
sentential connective connects two sentences:
John came to the party and Joe came to the party.
Then letting P be “John came to the party” and Q be “Joe came to the
party”, we can symbolize it as (P & Q).
Another example of such paraphrasing is found in the following
statement:
If either Kennedy or Khrushchev had been weaker willed concerning
either Berlin or Cuba, the cold war would have turned into a hot war.
This must be paraphrased first as
If either Kennedy had been weaker willed concerning either Berlin or
Cuba, or Khrushchev had been weaker willed concerning either Berlin
or Cuba, then the cold war would have turned into a hot war.
Then this must be paraphrased again as
If either Kennedy had been weaker willed concerning Berlin or Kennedy
had been weaker willed concerning Cuba, or Khrushchev had been
weaker willed concerning Berlin or Khrushchev had been weaker willed
concerning Cuba, then the cold war would have turned into a hot war.
Finally then, this can be symbolized as {[(P ∨ Q) ∨ (R ∨ S)] → T}.
Other kinds of paraphrasing may also be necessary. There are
expressions in English such as “unless”, “ but”, “if”, “only if”, “neither …
nor”, that are much like the sentential connectives. Whenever these occur
in a statement, the statement must be paraphrased to replace them by
sentential connectives. For example,
Neither Joe came to the party nor John came to the party
means the same thing as
Joe didn't come to the party and John didn't come to the party
and so it must be paraphrased in that way. In general, “Neither P nor Q”
can be paraphrased as (~P & ~Q). Equivalently, it can be symbolized as
~(P ∨ Q).
“But” is like “and” but emphasizes a contrast because the two conjuncts.
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 43

We may say “He came but he didn't like it” rather than “He came and he
didn't like it” merely to indicate that the two conjuncts do not go naturally
together. This emphasis makes no difference for logic, so we can symbolize
“P but Q “ as (P & Q).
The behavior of the expressions “if” and “only if” is somewhat
surprising. There is a strong temptation to identify “P if Q” with “If P then
Q”, but in fact it should be the other way around. That is, “P if Q” means
the same thing as “If Q then P”. Consider the statement “The crops will be
destroyed if there is a flood”. To say this is not to say that the crops might
not be destroyed anyway, for example, by a drought. But the statement “If
the crops are destroyed there is a flood” precludes the possibility of them
being destroyed by a drought as opposed to a flood. Therefore it cannot be
a proper paraphrase of “The crops will be destroyed if there is a flood”.
The proper paraphrase is “If there is a flood then the crops will be destroyed”.
In general, “P if Q” can be symbolized as (Q → P).
“P only if Q” works just the other way around. It means “If P then
Q”. Consider the statement “The crops will be destroyed only if there is a
flood”. This means that the only way the crops can be destroyed is by a
flood, and hence if the crops are destroyed then there must have been a
flood. This then means “ If the crops are destroyed then there is a flood”.
Note that “P if and only if Q” is just the conjunction of “P if Q” and “P only
if Q”, and thus that (P ↔ Q) is equivalent to [(P → Q) & (Q → P)].
One further expression that can be paraphrased in terms of the sentential
connectives is “unless”. “P unless Q” can be paraphrased as (~Q → P).
Suppose we want to paraphrase the statement “We will go to the beach
unless it rains”. This is the same thing as saying “If it doesn't rain then we
will go to the beach” that is (~Q → P).
One thing to beware of in symbolizing the forms of statements is that
there are other uses of some of the sentential connectives in which they do
not have their ordinary meaning. We noted in chapter one that “and”
sometimes means “and then” rather than simply “and”. In its ordinary use,
“P and Q” means “It is true that P, and it is true that Q”. On this reading,
“P and Q” means the same thing as “Q and P”. For example, “This is red
and that is white” means the same thing as “That is white and this is red”.
But consider the use of “and” in “He lay down and fell asleep”. This
clearly does not mean the same thing as “He fell asleep and lay down”.
The former means “He lay down and then fell asleep”. This temporal use
of “and” is not among the logical concepts the propositional calculus deals
with. In particular, it cannot be symbolized simply as “&”.

Exercises

Symbolize the forms of the following statements, using the sentential letters
indicated:
1. Neither Jack nor Jim will come unless Mary comes. (A I M)
2. We will not get there on time unless we speed, but if we speed we
will not get there at all. (T S G)
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 44

3. We can get the door open only if we use an acetylene torch on it, but
then the door will be ruined. (O A R)
4. The river will not overflow its banks unless we either have an early
thaw or heavy rains, but we will not have heavy rains. (O E H)
5. Unless we have a flat tire, we can get there on time if we speed, but
we will have a flat tire if we speed. (F T S)
6. Neither Jack nor Jim will come if Mary comes, unless Joan and Mary
both come. (A I M O)
7. Jeremy will get a Mercedes Benz for Christmas only if he does not
offend Santa Claus, but Jeremy will offend Santa Claus if he does not
believe in him, and Jeremy does not believe in Santa Claus. (M O B)
8. Rain is imminent. (you pick the sentential letters on this one)
9. John will not come unless Jim comes, and Jim will not come if Jeffrey
comes, but Jeffrey will only come if John does not come. (O I E)
10. It will rain if the barometer drops, but if it rains it will cool off later,
and it will not cool off later. (R B C)

7. Summary
The propositional calculus studies statement forms and argument forms
that can be constructed out of “and”, “or”, “it is false that”, “if ... then”, and
“if and only if”. These are the sentential connectives, and they are symbolized
using the logical symbols &, ∨, ~, →, and ↔.

Formulas of the propositional calculus were defined to be those expressions


that can be constructed using the six rules on page 29. For all but atomic
formulas and negations, formulas must have outer parentheses.

In symbolizing statement forms, it is best to begin with the smallest parts


and work outwards.

Some statements must be paraphrased before they can be symbolized:


“Neither P nor Q” is symbolized as ~(P ∨ Q) or as (~P & ~Q).
“P but Q” is symbolized as (P & Q).
“P if Q” is symbolized as (Q → P).
“P only if Q” is symbolized as (P → Q).
“P unless Q” is symbolized as (~Q → P).

Subjunctive conditionals are about what would be the case if something else
were the case.

Indicative conditionals are about what is the case if something else is the case.

The material conditional (P → Q) is defined to mean (~P ∨ Q). It is unclear


whether this is a proper symbolization of the indicative conditional.
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 45

Practice Problems

Symbolize the forms of the following statements, using the sentential letters
indicated:

1. If Meacham is impeached (I) or convicted (C), then unless he resigns


before he is removed from office (R), he will be unable to run in the
recall election (U).
2. If Robertson wins Florida (F) and Jackson wins Tennessee (T), then if
Hart loses in Georgia (G), Gephart will win in Indiana (I) if he is still
in the race (R).
3. If Victoria and the Kaiser were cousins (C), then Germany and England
were friendly (F), in which case World War II never happened (W).
4. If you take the test (T) and fail (F) you will be in trouble with the
university (U), but if you take the test and pass you will be in trouble
with the other students (S), and if you don’t take the test you will be
in trouble with the Dean (D).
5. If Henry is in pursuit of Oriental mysteries (P), then if he finds the
dragon lady (L) and the Foo Dog (D), he will be unable to emulate
Charlie Chan (C) unless either the Dragon Lady is benevolent (B) or
the Foo Dog is broken (K).
6. If the mining venture proves successful (S) but Jacob backs out too
early (E), then Jacob’s bank account will suffer (U) and his greedy
parents will disown him (D) unless he turns to a life of crime (C).
7. If the job is finished on time (F), we will be able to eat lunch on the
lawn (L), but if it isn’t finished on time we will have to eat lunch in
the cellar (C), and in that case we will have to share our lunch with
the mice (M).
8. Neither England nor France will be competitive in downhill skiing
(E,F) unless Switzerland beats Austria (B), but Switzerland won’t
beat Austria unless Schwartz breaks his leg (S).
9. If the plant will die (D) unless it gets water (W), but it won’t get
water if the sun shines (S) (unless there is a miracle (M)), then the
plant will die.
10. If Marilyn Monroe married Guy Lombardo (M) and they had a son
named “Ronnie” (R), he too could be president (P) if he moved to
Old Tucson (T) and became a cowboy movie star (C).
THE SYNTAX OF THE PROPOSITIONAL CALCULUS 46
3
The Semantics of the
Propositional Calculus

1. Interpretations
Formulas of the propositional calculus express statement forms. In
chapter two, we gave informal descriptions of the meanings of the logical
symbols of the propositional calculus, and relied upon that for our
understanding of the statement forms expressed by formulas. The next
step is to give a more precise description of those meanings. The objective
is to give a description that is sufficiently precise to allow us to use
mathematical tools in studying formal necessity and the validity of argument
forms.
Our general approach will be the same as for the implicational calculus.
We will begin by defining the notion of an interpretation, and then we will
give truth rules for the logical symbols. It is important to bear in mind that
formulas do not express statements—they express statement forms. As
such, they are not true or false. Typically, one and the same formula will be
the form of both true statements and false statements. For example, (P ∨ Q)
is the form of both of the following statements:

Either mice have whiskers or worms have wings.

Either mice have wings or worms have whiskers.

But the first statement is true and the second false.


This means that a truth value can only be associated with a formula
by first assigning meanings to its atomic parts. Once we do that, we pick
out a particular statement rather than just a statement form. We can say
that the statement form is true or false relative to an assignment of meanings if,
and only if, the statement that results from that assignment of meanings is
true or false. In other words, attributing truth values to formulas (relative
to interpretations) is just a shorthand way of talking about the truth values
of corresponding statements.
The most natural way to define the notion of an interpretation would
be to identify it with an assignment of statements to sentential letters. Once
we have said what statements the sentential letters stand for, it is determined
what statement is expressed by formulas built out of those sentential letters.
However, formulas of the propositional calculus have the nice characteristic
that we do not have to know their complete meanings to know their truth
values. For instance, consider a conjunction (P & Q). If we know that P and
Q are both true, then we can conclude that (P & Q) is true, and we can do
this without knowing what statements P and Q express. It will turn out
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 48

that this is true for all formulas of the propositional calculus. For this
reason, we will take an interpretation to simply assign truth values to
sentential letters rather than assign entire meanings. So:
An interpretation of the propositional calculus is an assignment of truth
values (truth or falsity) to the sentential letters.
We can then talk about formulas being true or false relative to an interpretation.

2. Truth Rules
To verify the claim that we can compute the truth value of any formula
if we know the truth values of the sentential letters occurring in it, we must
give the truth rules that allow us to make that computation. The rules we
employ are the following:

If A is any formula, ~A is true iff A is false.


If A and B are any formulas, ( A & B) is true if and only if both A and
B are true.
If A and B are any formulas, ( A ∨ B) is true if and only if either A or B
(or both) are true.
If A and B are any formulas, (A → B) is true if and only if either A is
false or B is true.
If A and B are any formulas, (A ↔ B) is true if and only A and B have
the same truth value, i.e., either both are true or both are false.
The truth rules for all but the conditional are self-explanatory, and the truth
rule for the conditional is just the definition of the material conditional, as
we discussed in chapter two.
Repeated application of these truth rules allows us to determine the
truth or falsity of complex formulas. To illustrate, suppose we are given
that P and Q are true, and R and S are false, and we want to determine
whether the formula {[(P & ~S) → ~Q] ↔ ~(R ∨ S)} is true. We begin with
the smallest parts. S is false, so ~S is true. Then P and ~S are both true, so
by Rule 2, (P & ~S) is true. Q is true, so by Rule 1, ~Q is false. Then by Rule
4, as (P & ~S) is true and ~Q is false, the conditional [(P & ~S) → ~Q] is
false. Neither R nor S is true, so by Rule 3, (R ∨ S) is false. Then by Rule 1,
~(R ∨ S) is true. But then the two sides of the biconditional
{[(P & ~S) → ~Q] ↔ ~(R ∨ S)} have different truth values (the left side is
false and the right side is true), and so, by Rule 5, the biconditional is false.
We can diagram this procedure as follows:
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 49

P true S false Q true R false S false

rule 1

~S true rule 1
rule 2

(P & ~S) true


~Q false rule 3

rule 4
[(P & ~S) → ~Q] false (R ∨ S) false

rule 1

~(R ∨ S) true

rule 5
{[(P & ~S) → ~Q] ↔ ~(R ∨ S)} false

Given the truth values of the atomic parts of a formula, it becomes a mechanical
matter to calculate the truth value of the whole formula using the above
truth rules. We begin with the smallest parts of the formula, calculate their
truth values, and then work outwards until we get the whole formula. The
calculation of the truth value parallels the construction we would use in
building the formula from its atomic parts.

3. Truth Tables
The truth rules allow us to compute the truth value of any formula
relative to an interpretation. There are infinitely many sentential letters,
because we allow the use of numerical subscripts. It follows that there are
infinitely many interpretations of the propositional calculus. However, when
focusing on a single formula, the only part of an interpretation that is relevant
to its truth value is the assignment to the sentential letters that are atomic
parts of the formula. There will be only finitely many ways interpretations
can assign truth values to any finite set of sentential letters. For example, if
we consider the sentential letters P, Q, and R, we can tabulate all the ways
of assigning truth values to them as follows:
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 50

P Q R

T T T
T T F
T F T
T F F
F T T
F T F
F F T
F F F

Note how we arranged the truth values to ensure that we have an exhaustive
list. In the rightmost column (under R), the truth values alternate. In the
next column (under Q) they alternate two at a time. In the final column
(under P) they alternate four at a time. If there were another column, they
would alternate eight at a time,n
and so on. In general, given a list of n
sentential letters, there are 2 combinations of truth values that can be
assigned to them. Thus for two sententential letters there are four
combinations of truth values, for three sentential letters there are eight, for
four there are sixteen, and so forth.
We can express the truth rules for the logical symbols in tabular form
as follows:

P ~P
T F
F T

P Q (P & Q)
T T T
T F F
F T F
F F F

P Q (P ∨ Q)
T T T
T F T
F T T
F F F

P Q (P → Q)
T T T
T F F
F T T
F F T
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 51

P Q (P ↔ Q)
T T T
T F F
F T F
F F T

In a similar way, we can tabulate the truth values of that formula for
all possible combinations of truth values of its atomic parts. Consider the
formula [(P & ~Q) ↔ (~(P ∨ Q) → Q)]. The following table can be constructed.
At each step the column of truth values being computed is greyed out, and
the columns from which the computations is made are marked in a lighter
grey:

Step 1
P Q [(P & ~Q) ↔ (~(P ∨ Q) → Q)]
T T
T F
F T
F F

Step 2 Truth values for ~Q


P Q [(P & ~Q) ↔ (~(P ∨ Q) → Q)]
T T F
T F T
F T F
F F T

Step 3 Truth values for (P & ~Q).


P Q [(P & ~Q) ↔ (~(P ∨ Q) → Q)]
T T F F
T F T T
F T F F
F F F T
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 52

Step 4 Truth values for (P ∨ Q)


P Q [(P & ~Q) ↔ (~(P ∨ Q) → Q)]
T T F F T
T F T T T
F T F F T
F F F T F

Step 5 Truth values for ~(P ∨ Q)


P Q [(P & ~Q) ↔ (~(P ∨ Q) → Q)]
T T F F F T
T F T T F T
F T F F F T
F F F T T F

Step 6 Truth values for (~(P ∨ Q) → Q)


P Q [(P & ~Q) ↔ (~(P ∨ Q) → Q)]
T T F F F T T
T F T T F T T
F T F F F T T
F F F T T F F

Step 7 Truth values for [(P & ~Q) ↔ (~(P ∨ Q) → Q)]


P Q [(P & ~Q) ↔ (~(P ∨ Q) → Q)]
T T F F F F T T
T F T T T F T T
F T F F F F T T
F F F T T T F F
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 53

This table is called a truth table. On the left-hand side of a truth table we list
all possible combinations of truth values for the atomic parts of the formula.
On the right-hand side are columns corresponding to each occurrence of a
logical symbol in the formula, and in that column we list the truth value of
the part of the formula that is constructed using that connective. We enclose
the column of truth values that are the truth values of the entire formula in
a box and print it in bold italic type.
The truth table for a formula gives us a record of its truth value under
all possible interpretations. The only thing difficult about constructing truth
tables is keeping track of which columns to use in computing the truth
values for each new column. It is easy to become confused about that. The
best way to avoid confusion is to keep in mind how the formula is constructed.
For example, in step 7 we are computing the truth value of a biconditional.
Ask yourself what the left and right sides of the biconditional are, and
which columns record their truth values. Those are then the columns to
which you appeal in computing the truths values of the biconditional. You
may find it useful to connect matching parentheses with lines, as was done
on page 33.

Exercises

Construct truth tables for the following formulas:


1. ~(P → ~~P)
2. [(P → Q) ∨ ~(P & ~Q)]
3. [(P ↔ Q) ↔ ~(P ↔ ~Q)]
4. [((P & Q) → R) & ~((R ∨ P) ↔ ~(P → Q))]
5. [ ~(P → Q) → ~([(R ∨ P) & ~(Q → ~P)] ↔ ~[(R & ~Q) → (P & ~R)])]
6. [ ~(P & Q) ↔ (~P & ~Q)]

4. Tautologies
Our main interest is in the formal necessity of statement forms and the
formal soundness of argument forms. Using truth tables we can give very
simple characterizations of these concepts for statement forms and argument
forms that can be symbolized in the propositional calculus. Let us begin by
discussing formal necessity.
Consider the logically necessary statement “Either it is snowing or it is
not snowing”. This statement has the form (P ∨ ~P). Any statement that
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 54

has this form will be necessary. Why is this so? Because regardless of what
we let P mean, either P or ~P will be true, and so (P ∨ ~P) will be true. To
make this clearer, consider the truth table for (P ∨ ~P):

P (P ∨ ~P)

T T F
F T T

Notice that (P ∨ ~P) comes out true on every line of this truth table. This is
enough to make it formally necessary, because regardless of how we interpret
P, it will have one of the truth values listed in the truth table, and then (P ∨
~P) will be true. Thus we can see by examining the truth table that any
statement having the form (P ∨ ~P) will be true under all circumstances,
regardless of whether P is true or false. This means that the statement is
necessary. So any statement of the form (P ∨ ~P) is necessary, and hence
this statement form is formally necessary.
It was remarked in chapter one that a formula true under every
interpretation is said to be valid. In the propositional calculus, a formula is
valid if, and only if, it is true on every line of its truth table. Such formulas
are called tautologies. (P ∨ ~P) is therefore a tautology. We write ¤ A to
indicate that a formula A is a tautology. For example, we can write
¤ (P ∨ ~P).
The significance of tautologies is that they are the formally necessary
statement forms of the propositional calculus. To see this, suppose we have
some formula of that propositional calculus that is a tautology, such as
{[P → (Q → R)] ↔ [(P & Q) → R]}. Consider any statement that has this
form. Then, even without knowing whether P, Q, and R are true, we can
verify that the statement itself is true just by seeing that the formula comes
out true on every line of its truth table. This is because, regardless of what
statements are symbolized by P, Q, and R, they will have some truth values,
and those truth values will correspond to a line of the truth table. The
statement symbolized by [P → (Q → R)] ↔ [(P & Q) → R]} will then be true
if and only if the formula is true on that line of its truth table. But the
formula is true on every line of its truth table, so it follows that the statement
will be true regardless of how the world might be (regardless of the truth
values of P, Q, and R). In other words, any statement of this form will be
necessary, and the statement form will be formally necessary. Therefore,
we can conclude:
If a formula is a tautology then the statement form it symbolizes is
formally necessary.
The converse of this principle is also true:
If a formula is a not tautology then the statement form it symbolizes is
not formally necessary.
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 55

This is because for every line of the truth table there will be some statement
having the form symbolized by the formula and having the truth value the
formula has on that line. To illustrate, consider the formula [P → (P & Q)].
This is not a tautology. It will be false when P is true but Q is false. But
now pick any statements having the truth values that, when assigned to P
and Q, make [P → (P & Q)] false. For instance, we might let P stand for
“2+2 = 4” and Q stand for “2+3 = 6”. The statement
[2+2 = 4 → (2+2 = 4 & 2+3 = 6)]
is then be true if and only if [P → (P & Q)] is true on that line of its truth
table where P is true and Q is false. Hence the statement is false. It follows
that not all statements of the form [P → (P & Q)] are logically necessary (in
particular, this one isn’t), and so [P → (P & Q)] does not symbolize a
formally necessary statement form.
We can conclude in general then that:
A formula of the propositional calculus is a tautology if, and only if, it
is the form of a formally necessary statement form.
Thus truth tables and tautologies give us a way of talking about the formal
necessity of any statement whose form can be expressed in the propositional
calculus.

5. Metatheorems
Logical necessity was explained using metaphors and by giving
examples, but not by giving a precise definition. That has the consequence
that although we can reason about the properties of logical necessity and
the attendant concepts of formal necessity and the validity of argument
forms, we cannot literally prove theorems about these concepts in the sense
that one proves theorems in mathematics. In this respect, they contrast
sharply with the concept of a tautology. The concept of a tautology is a
mathematically precise concept. We could, for example, write a computer
program that mechanically checks formulas to determine whether they are
tautologies. It would do this by constructing a truth table.
Because tautologicity is a mathematically precise concept, we can
literally prove mathematical theorems about it. For example, one important
fact about tautologies is:
The conjunction of two tautologies is itself a tautology.
That is, if we begin with two formulas, A and B, and ¤ A and ¤ B, then
¤ (A & B). This principle is known as adjunctivity. To establish this,
suppose that ¤ A and ¤ B. Then any assignment of truth values to the
atomic parts of A and B will make them both true. But then any such
assignment of truth values will also make the conjunction (A & B) true, and
so it is also a tautology.
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 56

The preceding reasoning is simple, but the result is quite literally a


mathematical theorem about tautologicity. Theorems about a logical theory
are called metatheorems. The investigation of a logical theory proceeds largely
by proving metatheorems. It is the metatheorems that enable us to determine
the properties of the theory and investigate how, for example, to construct
arguments using formulas of the theory. Metatheorems need not be very
deep, and their proofs are often easy. This is illustrated by the proof of
adjunctivity. But even simple metatheorems can be important.
We have two classes of concepts pertaining to the propositional calculus.
We started off with the philosophical concepts of logical necessity, validity,
formal necessity, and the validity of argument forms. Then we introduced
the mathematical concept of tautologicity, and we will use it to define other
mathematical concepts. The philosophical concepts represent our initial
interest in the propositional calculus, but the mathematical concepts provide
a tool for studying the philosophical concepts. This is made possible by the
argument given in the previous section to the effect that for formulas of the
propositional calculus, tautologicity and formal necessity coincide. Thus
we can study the philosophically interesting concept of formal necessity by
studying the mathematical concept of tautologicity, and we can use all of
the tools of mathematics in studying the latter. The mathematics required
need not be very difficult, but what it achieves is rigorous proofs of precise
results.
By proving metatheorems, we become clearer on the logical structure
of the propositional calculus, and this will eventually enable us to prove
important results about how to reason using formulas of the propositional
calculus. The metatheorems comprise what is called the metatheory of the
propositional calculus. The metatheory is just the collection of all our
metatheorems.
To advance our study of the propositional calculus, we can define two
concepts that are closely related to tautologicity. We say that a formula is
truth-functionally consistent if, and only if, it is true on at least one line of its
truth table. A formula is truth-functionally inconsistent if, and only if, it is
not truth-functionally consistent—that is, if and only if it is false on every
line of its truth table. Because these concepts are defined in terms of
tautologicity, they are also mathematically precise and we can prove
metatheorems about them. For instance, here is a simple metatheorem
about truth-functional inconsistency:
Metatheorem: A formula is truth-functionally inconsistent if, and only
if, its negation is a tautology.
Proof: A truth-functionally inconsistent formula is one that is false on every
line of its truth table. A formula is false on every line of its truth table if,
and only if, its negation is true on every line of its truth table. But the negation
is true on every line of its truth table if, and only if, it is a tautology. ■

Note that “■” marks the end of the proof.


We can prove many similar metatheorems by thinking about truth
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 57

tables and assignments of truth values to the atomic parts of formulas.


Usually, such a proof begins by “expanding the definitions” of the terms
employed in the metatheorem. For example, the above proof began by
expanding the definitions of truth-functional inconsistency and tautologicity.
Then we reason about the conditions that are stated in the expanded form
of the principle being proven. The student should keep this general strategy
in mind in doing the following exercises.

Exercises

A. Prove the following metatheorems by giving arguments analogous to


those given in the text.
1. Every tautology is truth-functionally consistent.
2. If one formula is a tautology, and a second formula is truth-functionally
consistent, then their conjunction is truth-functionally consistent.
B. To prove that something holds in general one must give a general
argument. However, in the following exercises, you are asked to prove that
something is possible. This can be done by giving a single example.
1. It is possible for two formulas to be each truth-functionally consistent
while their conjunction is truth-functionally inconsistent.
2. It is possible for both a formula and its negation to be truth-functionally
consistent.

6. Tautological Implication
The concept of tautologicity provides the vehicle for a mathematical
study of formal necessity in the propositional calculus. We can define an
analogous concept that allows us to study the validity of argument forms
expressed in the propositional calculus. We saw in chapter one that an
argument form is valid if, and only if, its corresponding conditional is
formally necessary. In the propositional calculus, formal necessity coincides
with tautologicity, so it follows that:
An argument form expressed in the propositional calculus is valid if,
and only if, its corresponding condition is a tautology.
This is captured by the concept of tautological implication. We define:
A set of formulas A1,…,An tautologically implies a formula B if, and
only if, ¤ [(A1 &…&An) → B].
We abbreviate “A1,…,An tautologically implies B” as “A1 ,…,An ¤ B”.
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 58

The significance of tautological implication is given by the following


principle:
An argument form expressed in the propositional calculus is valid if,
and only if, its premises tautologically imply its conclusion.
Tautological implication consists of the tautologicity of the corresponding
conditional, and that can be checked by constructing a truth table. So given
an argument form expressed in the propositional calculus, we can assess its
validity by constructing a truth table and determining whether its
corresponding conditional is a tautology.
A conditional is a tautology just in case its conclusion is true on every
line of its truth table on which its antecedent is true. Thus an argument
form expressed in the propositional calculus is valid (its premises
tautologically imply its conclusion) if, and only if, every line of its truth
table on which its premises are true is also a line on which its conclusion is
true. Hence truth tables become a valuable tool for investigating the validity
of argument forms. It has to be emphasized, however, that the scope of this
tool is limited to argument forms that can be expressed in the propositional
calculus. There are many valid argument forms that cannot be expressed in
the propositional calculus. For example:

All A are B.
All B are C.

Therefore, all A are C

is a valid argument form, but it cannot be expressed in the propositional


calculus. To deal with arguments like this we need the greater expressive
power of the predicate calculus, which will be investigated in part II of the
book. These arguments cannot be assessed using truth tables.
It will be useful to have a stock of tautological implications to which
we can appeal when we begin investigating how to reason and construct
arguments within the propositional calculus. The following tautological
implications are important and are easily verified using truth tables:

I1. (A & B) ¤ A 
 simplification
I2. (A & B) ¤ B 
I3. A ¤ (A ∨ B) 
 addition
I4. B ¤ (A ∨ B) 
I5. ~A ¤ (A → B)
I6. B ¤ (A → B)
I7. ~(A → B) ¤ A
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 59

I8. ~(A → B) ¤ ~B
I9. A, (A → B) ¤ B modus ponens
I10. ~B, (A → B) ¤ ~A modus tollens
I11. ~A, (A ∨ B) ¤ B 
 disjunctive syllogism
I12. ~B, (A ∨ B) ¤ A 
I13. (A → B), (B → C) ¤ (A → C) hypothetical syllogism
I14. A, B ¤ (A & B) adjunction
I15. (A ∨ B), (A → C ), (B → C) ¤ C dilemma

Let us see why these implications hold:


I1 and I2 both hold because a conjunction is true if and only if both conjuncts
are true.
I3 and I4 hold because a disjunction is true if either disjunct is true.
I5 holds because a conditional is true if its antecedent is false.
I6 holds because a conditional is true if its consequent is true.
I7 and I8 hold because a conditional is only false when its antecedent is true
and its consequent false.
I9: Suppose A is true and (A → B) is true. Then we have a true conditional
having a true antecedent. But then, in order for the conditional to be true,
the consequent must also be true. Thus B must be true.
I10: Exercise for the student.
I11: Suppose ~A and (A ∨ B) are both true. In order for the disjunction to
be true, at least one of its disjuncts must be true. But the first disjunct is
false, because ~A is true. Thus the second disjunct, B, must be true. I12 is
analogous.
I13: Exercise for the student.
I14: If both A and B are true, then both conjuncts of (A & B) are true, so
the conjunction is true. Note that although this principle is called adjunction,
it is a different principle from the adjunctivity of tautologicity. This principle
is the adjunctivity of tautological implication.
I15: Suppose (A ∨ B), (A → C ), and (B → C ) are all true. Since the
disjunction (A ∨ B) is true, either A is true or B is true. But if A is true then
C is true (by the first conditional), and if B is true then C is true (by the
second conditional). Thus if either A or B is true, then C is true. But we
know that either A is true or B is true, so C is true.
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 60

Exercises

Prove that I10 and I13 hold

7. Tautological Equivalence
On several occasions it has been remarked that one formula is
“equivalent” to another. For instance, it was remarked that “Neither P nor
Q” could be symbolized as either (~P & ~Q) or ~(P ∨ Q), and that the two
formulas were equivalent. Now we are in a position to make this concept
of equivalence precise. The sense in which two formulas of the propositional
calculus can be equivalent is that they are true under the same circumstances.
More precisely, they are true relative to the same interpretations. This
concept of equivalence is called tautological equivalence:
Two formulas of the propositional calculus are tautologically equivalent
if, and only if, they are true relative to the same interpretations.
Thus, for example, (~P & ~Q) and ~(P ∨ Q) are true relative to the same
interpretations. This can be verified by constructing a truth table:
P Q (~P & ~Q) ~(P ∨ Q)
T T F F F F T
T F F F T F T
F T T F F F T
F F T T T T F

If two formulas have the same truth values, then their biconditional is
true. E.g., we could rewrite the preceding truth table as follows:
P Q [(~P & ~Q) ↔ ~(P ∨ Q)]
T T F F F T F T
T F F F T T F T
F T T F F T F T
F F T T T T T F

Thus we could define tautological equivalence just as well as follows:


THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 61

Two formulas of the propositional calculus, A and B, are tautologically


equivalent if, and only if, (A ↔ B) is a tautology.
We will often abbreviate “tautologically equivalent” as “eq.” The
following tautological equivalences can all be verified very easily using
truth tables:

E1. ~~A eq. A double negation


E2. ~(A ∨ B) eq. (~A & ~ B) 
 De Morgan's Laws
E3. ~(A & B) eq. (~A ∨ ~B) 
E4. (A → B) eq. (~A ∨ B)
E5. ~(A → B) eq. ( A & ~B)
E6. ~(A ↔ B) eq. (A ↔ ~B)
E7. (A ↔ B) eq. [(A → B) & (B → A)]
E8. (A ↔ B) eq. [(A & B) ∨ (~A & ~B)]
E9. (A & ( B & C )) eq. ((A & B) & C ) 
 associative laws
E10. (A ∨ (B ∨ C)) eq. (( A ∨ B) ∨ C ) 
E11. (A & B) eq. (B & A) 
 commutative laws
E12. (A ∨ B) eq. ( B ∨ A) 
E13. [A & ( B ∨ C)] eq. [( A & B) ∨ (A & C)] 
 distributive laws
E14. [A ∨ (B & C)] eq. [( A ∨ B) & (A ∨ C )] 
E15. [A → (B → C )] eq. [(A & B) → C ] exportation-importation
E16. (A → B) eq. (~B → ~ A) contraposition
E17. (A ∨ A) eq. A 
 idempotency laws
E18. (A & A) eq. A 
E19. (A ∨ B) eq. (~A → B)
E20. (A ∨ B) eq. (~B → A)

Let us examine each of these equivalences separately, and see why it is true:
E1. ~~A is true if, and only if, ~A is not true, and ~A is not true if, and
only if, it is not the case that A is not true; that is, if and only if A is true.
E2. ~(A ∨ B) is true if, and only if, ( A ∨ B) is false. But (A ∨ B) is false if,
and only if, both disjuncts are false; that is, if and only if both A and B are
false. But this is the same as saying that ~A and ~B are both true; that is,
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 62

that (~A & ~B) is true.


E3. ~(A & B) is true if, and only if, (A & B) is false. But (A & B) is false
if, and only if, at least one conjunct is false; that is, if and only if either A is
false or B is false. But this is the same as saying that either ~A is true or ~B
is true; that is, that (~A ∨ ~ B) is true.
E4. We saw in Section 4 that (A → B) is true if, and only if, (~ A ∨ B) is
true.
E5. ~(A → B) is true if, and only if, (A → B) is false. (A → B) is false if,
and only if, the antecedent is true and the consequent is false, that is, A and
~B are true. Thus (A → B) is false if, and only if, (A & ~B) is true. Thus,
~(A → B) is true if, and only if, ( A & ~B) is true.
E6. ~(A ↔ B) is true if, and only if, A and B have opposite truth values;
that is, if and only if A and ~B have the same truth values; that is, if and
only if (A ↔ ~B) is true.
E7. (A ↔ B) is true if, and only if, A and B have the same truth value.
Thus if A is true then B must be true, and if B is true then A must be true.
So if (A ↔ B) is true, then
[(A → B) & (B → A)]
must be true. Conversely, if [(A → B) & (B → A)] is true, then if A is true,
so must B be true, and if A is false, so must B be false. Thus ( A ↔ B) is
true. Therefore, (A ↔ B) is true if, and only if, [(A → B) & (B → A)] is
true.
E8. Exercise for the student.
E9 is obvious.
E10 is obvious.
E11 is obvious.
E12 is obvious.
E13. To say that [A & ( B ∨ C )] is true is to say that both A and (B ∨ C) are
true. Thus A must be true, and either B or C must be true. So either (A &
B) will be true, or (A & C ) will be true. So if [A & (B ∨ C )] is true, then [(A
& B) ∨ (A & C )] is true.
Conversely, if [(A & B) ∨ (A & C)] is true, then either (A & B) or (A &
C) must be true. In either case, A must be true. In addition, in order to
have one of (A & B) or (A & C ) true, we must have either B true or C true.
So (B ∨ C) must be true. Thus both A and ( B ∨ C) must be true, so [ A & (B
∨ C )] must be true. Thus [A & (B ∨ C)] is true if, and only if, [(A & B) ∨ (A
& C)] is true.
E14. Verify this with a truth table.
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 63

E15. Exercise for the student.


E16. If (A → B) is true, then if B is false, A must also be false, because if A
were true, then the conditional would have a true antecedent and false
consequent, which would make it false. Thus if ~B is true, then ~A is true;
that is, (~B → ~A) is true.
Conversely, if (~B → ~ A) is true, then by the same argument, (~~A →
~~B) must be true. But ~~A is equivalent to A, and ~~B is equivalent to
B, so (A → B) must be true.
E17 is obvious.
E18 is obvious.
E19. By E1, (A ∨ B) is equivalent to (~~ A ∨ B), and by E4 this is equivalent
to (~A → B).
E20 is analogous to E19.

Exercises

A. Given arguments showing that E8 and E15 hold.

B. Prove the following metatheorem:


Two formulas of the propositional calculus are tautologically equivalent
if, and only if, they tautologically imply each other.

8. Some Metatheorems

Tautologicity, tautological implication, and tautological equivalence


are the main semantical concepts of the propositional calculus. They have
each received characterizations in terms of truth tables. In principle, any
claim of tautologicity, tautological implication, or tautological equivalence
can be verified by constructing a truth table. However, in practice that is
often too difficult. For example, a truth table for an implication involving 8
atomic parts would require a truth table with 256 lines. It would be impractical
to try to construct such a truth table. An implication involving 16 atomic
parts would require a truth table with 65536 lines. A computer could
construct such a truth table, but a human being would probably get it
wrong. An implication involving 300 atomic parts would require a truth
90
table with approximately 10 lines. Not even a computer could construct
78
that truth table. It has been estimated that there are 10 elementary particles
in the universe, so the truth table would have more lines than there are
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 64

particles in the universe. And yet, it is not out of the question that we could
show such an implication to hold. For example, it might have the form
(P1 & ... & P300 ) ¤ P1 .
We can give a very simple argument to show that this implication holds,
without having to construct a truth table.
The preceding example shows that it is often better to reason about
tautologicity and tautological implication rather than constructing truth
tables. Reasoning about these concepts consists of proving metatheorems.
So let us turn to some metatheorems that will subsequently prove useful.
Let us begin with a simple one:
Metatheorem: If a formula is either tautologically equivalent to, or
tautologically implied by a tautology, then it is itself a tautology.
Proof: This holds because, if if a formula is tautologically implied by a
tautology, then the formula must be true on every line of its truth table on
which the tautology is true. Since the tautology is true on every line of the
truth table, the formula itself must also be true on every line of its truth
table; thus it is a tautology. ■
Next note that tautological implication is transitive:
Metatheorem: If one formula (or set of formulas) tautologically implies
a second formula, and the second formula tautologically implies a third,
then the first formula (or set of formulas) tautologically implies the
third.
Proof: Suppose we have a set of formulas A1 ,…,An and B and C , and
suppose A1 ,…,An ¤ B and B ¤ C . Let us show that it follows from this that
A1 ,…,An ¤ C . By the definition of tautological implication, as A1 ,…,An ¤ B
and B ¤ C , C is true on every line of the truth table on which B is true, and
B is true on every line on which A1 ,…,An are true. Hence C is true on every
line on which A1,…,An are true. That means that A1,…,An ¤ C. ■
Because tautological implication is transitive, we can establish
tautological implications using several steps. For example, suppose we
want to show that (P & Q) ¤ (P ∨ Q). We could do this by using a truth
table and showing that ¤ [(P & Q) → (P ∨ Q)]. But it is much easier to do it
as follows. By I1, (P & Q) ¤ P. By I3, P ¤ (P ∨ Q). Therefore, by the
transitivity of implication, (P & Q) ¤ (P ∨ Q).
Using the transitivity of tautological implication in this way we can
string simple implications together to establish more complicated
implications. If we want to show that one formula, A, implies another
formula, B, we might do this by showing that A implies some other formula
and that the second formula implies a third formula (and hence A implies
the third formula); then the third formula implies a fourth formula (and
hence A implies the fourth formula), and so on until we obtain the formula
B. In other words, if we have a string of implications of the form A ¤ A1 ,
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 65

A1 ¤ A2 , A2 ¤ A3 ,…, An ¤ B, it follows that A ¤ B. This is because, as A ¤


A1 , and A1 ¤ A2, we must have A ¤ A2, and then as A2 ¤ A3 , we must have
A ¤ A3 , and so on until we establish that A ¤ B.
Notice that in constructing strings of implications like this, we can
also use tautological equivalences, because if two formulas are tautologically
equivalent, then each tautologically implies the other. If we want to show
that (P ↔ Q) ¤ (~P → ~Q), we may use E7 to observe that
(P ↔ Q) ¤ [(P → Q) & (Q → P)].
Then by I2,
[(P → Q) & (Q → P)] ¤ (Q → P)
and by E16,
(Q → P) ¤ (~P → ~Q).
Finally, by the transitivity of implication,
(P ↔ Q) ¤ (~P → ~Q).
Given that tautological implication is transitive, it is easy to see that
tautological equivalence is also transitive:
Metatheorem: If one formula is tautologically equivalent to a second,
and the second is tautologically equivalent to a third, then the first is
tautologically equivalent to the third.
Proof: Suppose that A eq. B and B eq. C Then A ¤ B and B ¤ C , and so by
the transitivity of tautological implication, A ¤ C . Also B ¤ A and C ¤ B,
and so once again by the transitivity of tautological implication, C ¤ A.
Then because both A ¤ C and C ¤ A, we can conclude that A eq. C . ■
A further important fact about tautological implication is that it is
adjunctive:
Metatheorem: Given any three formulas, A, B, and C, if A ¤ B and A ¤ C ,
then A ¤ (B & C).
Proof: In other words, if a formula implies each of two other formulas, then
it implies their conjunction. To see that this is true, suppose A ¤ B and A ¤
C. Any assignment of truth values to the atomic parts of A, B, and C that
makes A true will make both B and C true, so any such assignment will
make (B & C) true, and hence A ¤ (B & C). ■
To illustrate the use of this metatheorem, suppose we want to show
that P, (P ↔ Q) ¤ Q without constructing a truth table. Clearly
P, (P ↔ Q) ¤ P
and
P, (P ↔ Q) ¤ (P ↔ Q).
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 66

By E7,
(P ↔ Q) ¤ [(P → Q) & (Q → P)],
and by I1,
[(P → Q) & (Q → P)] ¤ (P → Q).
Thus by the transitivity of implication,
P, (P ↔ Q) ¤ (P → Q).
Then by the fact that implication is adjunctive,
P, (P ↔ Q) ¤ [P & (P → Q)].
By I9,
[P & (P → Q)] ¤ Q.
Thus by transitivity again,
P, (P ↔ Q) ¤ Q.
We can generalize the adjunctivity of tautological implication
somewhat:
Metatheorem: If a set of formulas A1 ,…,An tautologically implies each
formula in another set B1 ,…,Bm then A1,…,An implies the conjunction
(B1 & … & Bm) (where the inner parentheses can be in any order).
Proof: Suppose the set A1,…,An implies each of B1 ,…,Bm. Then as A1,…,An
¤ B1 and A1,…,An ¤ B2, we must have A1,…,An ¤ (B1 & B2). Then as
A1 ,…,An ¤ B3 we must have A1 ,…,An ¤ (B1 & B2 & B3). And so on. Thus
A1 ,…,An ¤ (B1 & … & Bm). ■
We can also generalize the transitivity of tautological implication
somewhat:
Metatheorem: If each formula in a set B1,…,Bm of formulas is
tautologically implied by the set of formulas A1,…,An. and B1,…,Bm ¤
C, then A1,…,An ¤ C .
This is called strong transitivity. The form of this principle becomes more
obvious if we diagram the relations between the formulas as follows:
A1 ,…,An ¤ B1 

A1 ,…,An ¤ B2 
 ¤C
... 

A1 ,…,An ¤ Bm 

Proof: To establish strong transitivity, suppose each of B1 ,…,Bm are implied


THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 67

by A1,…,An. Then as we have just seen, A1,…,An ¤ (B1 & … & Bm). If
B1 ,…,Bm ¤ C , then (B1 & … & Bm ) ¤ C . So by transitivity, A1 ,…,An ¤ C .
The difference between transitivity and strong transitivity is that in the
former we have a set of formulas implying a single formula, which in turn
implies another formula, whereas in the latter we have a set of formulas
implying a second set of formulas, which in turn implies another formula. ■
Strong transitivity is of fundamental importance in understanding the
structure of reasoning. Arguments typically have a kind of “tree structure”
wherein we start with some premises, draw intermediate conclusions from
the premises, draw further intermediate conclusions from the intermediate
conclusions, and finally infer our desired conclusion from some of those
intermediate steps. For instance, we might have an argument whose structure
could be diagrammed as follows:
A1 A2 A3 A4 A5

B1 B2

B3 B4

C
When we reason like this, we assume that the argument has established
that the final conclusion is implied by the initial premises. It is strong
transitivity that allows us to conclude that B3 and B4 are implied by the
initial premises, and then to conclude from that that C is implied by the
initial premises. Without strong transitivity, we would have to somehow
get the conclusion from the premises in a single step.
Another important fact about implication is the following:
Metatheorem: If C 1,…,C n, A ¤ B, then C1 ,…,Cn ¤ (A → B).
In other words, we can remove one formula from the head of the implication
and make it the antecedent of a conditional in the conclusion of the implication.
This is the principle of conditionalization. It is true for the following reason.
Proof: Suppose C 1,…,C n, A ¤ B. By the definition of tautological implication,
this means that the conditional [(C1 & … & C n & A) → B] is a tautology. By
E15 this is tautologically equivalent to the conditional
[(C 1 & … & C n) → (A → B)].
We have seen that a formula tautologically equivalent to a tautology is itself
a tautology, so ¤ (C1 & … & Cn) → (A → B)]. Then C1 ,…,Cn ¤ (A → B). ■
The principle of conditionalization is of considerable use in establishing
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 68

tautological implications. For example, suppose we want to show that


P ¤ [Q → (P & Q)].
By I14,
P, Q ¤ (P & Q).
Then by the principle of conditionalization,
P ¤ [Q → (P & Q)].
To take another example, suppose we want to show that
(P → R) ¤ [(P ∨ Q) → [(Q → R) → R]].
By I15,
(P ∨ Q), (P → R), (Q → R) ¤ R.
By the principle of conditionalization,
(P ∨ Q), (P → R) ¤ [(Q → R) → R].
Then by a second application of the principle of conditionalization,
(P → R) ¤ [(P ∨ Q) → [(Q → R) → R]].
Combining these facts about tautological implication enables us to
give arguments showing that one formula or set of formulas implies another
formula. For example, suppose we want to show that
(~P ∨ Q), (~Q ∨ R) ¤ (~P ∨ R).
We might reason as follows. Clearly
(~P ∨ Q), (~Q ∨ R) ¤ (~P ∨ Q)
and
(~P ∨ Q), (Q ∨ R) ¤ (~Q ∨ R).
By E4,
(~P ∨ Q) ¤ (P → Q),
and
(~Q ∨ R) ¤ (Q → R).
So by transitivity,
(~P ∨ Q), (~Q ∨ R) ¤ (P → Q)
and
(~P ∨ Q), (~Q ∨ R) ¤ (Q → R).
By I13,
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 69

(P → Q), (Q → R) ¤ (P → R).
Then by strong transitivity,
(~P ∨ Q), (~Q ∨ R) ¤ (P → R).
Byy E4,
(P → R) ¤ (~P ∨ R).
Then by transitivity,
(~P ∨ Q), (~Q ∨ R) ¤ (~P ∨ R).
General arguments like this provide an alternative to truth tables in
establishing tautologies and tautological implications. However, it can
reasonably be doubted whether the construction of such arguments is easier
than constructing a truth table. At this point the reader is no doubt wondering
how we know which principle to use at any given point in constructing a
proof. This question will be answered in the next chapter.

Exercises

Prove the following metatheorem:


1. A truth-functionally inconsistent formula tautologically implies every
formula.
2. A tautology is tautologically implied by every formula.
3. Any two tautologies are tautologically equivalent.

9. Failures of Tautologicity
To show that a formula is a tautology, one must either construct an
entire truth table or a give a general argument to that effect. This can be
difficult. It is generally much easier to show that a formula is not a tautology.
It suffices to find a single assignment of truth values (an interpretation) that
makes the formula false. Equivalently, one need only find a single line of
the truth table on which the formula is false. For example, suppose we
want to show that [(P → Q) & (Q → P)] is not a tautology. This is a
conjunction, so it suffices to make one conjunct false. For each conjunct,
there is only one way to make it false. For example, focusing on the first
conjunct, we can make P true and Q false. That makes the whole formula
false, showing that it is not a tautology.
A similar technique can be used to show that a proposed tautological
THE SEMANTICS OF THE PROPOSITIONAL CALCULUS 70

implication fails. For example, to show that the two premises (P → ~Q), (Q
→ ~R) do not jointly imply (P → R), it suffices to find an assignment of
truth values making the premises true and the conclusion false. The only
way to make the conclusion false is to make P true and R false. That
automatically makes the first premise true, and if we also make Q false then
the second premise is true as well. So the proposed implication does not
hold.

Exercises

Find interpretations showing the following:


1. [P ↔ (Q ↔ P)] is not a tautology.
2. [P ↔ (Q ↔ P)] does not tautologically imply P.
3. [P ∨ (Q & ~R)] does not tautologically imply (P ∨ R).
4
Derivations in the
Propositional Calculus

1. Arguments Expressed in
the Propositional Calculus
We have seen that we can symbolize a wide variety of statement
forms using formulas of the propositional calculus. It follows that we can
also express argument forms constructed out of those statement forms. Thus
when presented with the following argument:
1. If Joseph lives in Phoenix, he lives in Arizona.
2. If Joseph lives in Tucson, he lives in Arizona.
3. Joseph lives in either Phoenix or Tucson.
4. Therefore, Joseph lives in Arizona.
we might symbolize its form as follows:
(P → A)
(T → A)
(P ∨ T)
A
We have seen that an argument form expressed in the propositional calculus
is formally valid if, and only if, its premises tautologically imply its conclusion.
Equivalently, it is formally valid if, and only if, its corresponding conditional
is a tautology. Thus we could assess the formal validity of this argument
form by constructing a truth table for the formula
{[(P → A) & (T → A)] & (P ∨ T)] → A}.
However, there is a simpler way to go about it. We can observe that this
argument is an instance of I15, dilemma. Thus we already know that this is
a formally valid argument form.
Consider an argument that is a little more complicated:
1. If Joseph lives in either Arizona or California, he must pay sales tax
on this purchase.
2 If Joseph lives in Phoenix, he lives in Arizona.
3 If Joseph lives in Tucson, he lives in Arizona.
4. Joseph lives in either Phoenix or Tucson.
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 72

5. Therefore, Joseph must pay sales tax on this purchase.


This can be symbolized as follows:
[(A ∨ C) → S]
(P → A)
(T → A)
(P ∨ T)
S
This argument form is not an instance of any of the tautological implications
we have already noted. However, it can be obtained by stringing three
tautological implications together, using the principle of strong transitivity
discussed in chapter three. We can diagram this as follows:
(P ∨ T) (P → A) (T → A)
by I15
A
by I3

(A ∨ C) [(A ∨ C) → S]
by I9
S
This diagram is just a way of representing reasoning, using tautological
implications we already know to establish new ones. To make it look more
like ordinary reasoning, we can rewrite it as a sequence of formulas rather
than a tree:
1. [(A ∨ C) → S] given
2. (P → A) given
3. (T → A) given
4. (P ∨ T) given
5. A by I15, from (2), (3), and (4).
6. (A ∨ C) by I3 from (5)
7. S by I9 from (1) and (6).
What this illustrates is that much ordinary reasoning is just a way of
establishing tautological implications by using “obvious” implications and
principles like strong transitivity that tell us how new implications can be
obtained from old.
The purpose of this chapter is to investigate the process of constructing
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 73

such arguments in the propositional calculus. We will proceed by defining


the structure of such arguments more carefully, and investigating how one
can go about finding arguments to show that a particular conclusion follows
from a given set of premises. The arguments constructed in this way will
have a precisely delineated structure, and will be called derivations in the
propositional calculus.

2. Derivations
Derivations will be sequences of “lines”, where a line has four parts.
The general form of a line is:
(i,j,k,...) n. formula explanation
Lines are numbered sequentially, and n is the line number. Line numbers
are used for easy reference to particular lines. formula is a formula of the
propositional calculus. It is the formula inferred on that line of the derivation.
explanation is an explanation of how formula was inferred. Explanations will
always have the general form:
rule, from (a), (b), ...
where rule is a rule of inference and (a), (b), ... are the numbers of the lines
from which formula was inferred using the rule of inference.
An argument always begins with some premises, and what it establishes
is that its conclusion follows from its premises. So what a derivation is
really establishing is a tautological implication. Some of the lines of a
derivation represent premises, and other lines represent conclusions drawn
from those premises. Not all lines need be inferred from the same premises,
so we need a way of keeping track of which premises are used in getting
each line. That is the purpose of the initial list of numbers (i,j,k,...). These
are the line numbers of the premises used in getting that line, and are called
the premise numbers of the line.
The lines of a derivation are really shorthand for reports of tautological
implications. For instance, if a derivation contains a line
(3,5) 7. (P ∨ ~R) explanation
what this is telling us is that (P ∨ ~R) is tautologically implied by the
formulas on lines (3) and (5) of the derivation.
It will turn out that it is possible to have a line of a derivation that
lacks premise numbers. What that signifies is that the formula on that line
was inferred in such a way that it does not depend on any premises. In
other words, it is a tautology. Given that this is the significance of a line’s
not having premise numbers, all other lines must have premise numbers.
In particular, premises themselves must have premise numbers. This may
seem odd, because a premise does not depend upon anything else. It is just
a premise. What can we use for the premise numbers of such a line? A
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 74

simple choice is the line number of the line itself. That is, if we want to
record a premise P on, say, the third line of a derivation, we can do it by
writing:
(3) 3. P premise
This says that P is adopted as a premise on line (3), and the premise number
tells us that the formula on line (3) implies itself. This is a trivial use of
premise numbers, but its purpose is to reserve lines with no premise numbers
for recording tautologies.

3. Linear Derivations
To say precisely which sequences of lines count as derivations, we
must say which rules of inference we are permitted to use in constructing
derivations. Perhaps the simplest and most obvious rule of inference is one
telling us that if a formula A is tautologically implied by some other formulas
B1 ,…,Bm that already occur in the derivation, then we can write A on any
subsequent line, citing the lines on which B1 ,…,Bm occur in our explanation.
When we make an inference of this sort, we must have some fixed list of
tautologically implication to which we can appeal, and for this purpose we
will use I1–I15 and E1–E20 from chapter three. Equivalences are implications
that go in both directions, and so they can be used to make an inference in
either direction. For instance, we could cite E2 to make an inference from
~(P ∨ Q) to (~P & ~Q) or to make an inference from (~P & ~Q) to ~(P ∨ Q).
This is made precise by the following rule of inference:
RULE I: IMPLICATION: If, according to one of the implications I1 – I15 or
E1–E20, a set of formulas B1 ,…,Bm appearing on one or more lines
of a derivation tautologically implies a further formula A, then we
can write A on any later line of the derivation, taking as premise
numbers all of the premise numbers of B1,…,Bm.
Rule I allows us to add a line to a derivation by appealing to previous
lines, but to get started we need a rule for introducing premises. For this
purpose we will use the following rule:
RULE P: PREMISE INTRODUCTION: Any formula can be written on any line
of a derivation provided that we let the premise number be the line
number of that line.
This rule lets us take anything as a premise at any time. This may seem
overly liberal. Shouldn’t we only be allowed to take something as a premise
if it is known to be true? If our purpose is to establish the truth of the
conclusion, then there is no point in inferring it from premises not known
to be true, because all the derivation will establish is that the conclusion is
true if the premises are true. However, there is nothing logically wrong with
taking premises that are not known to be true. Recall that a line of a
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 75

derivation is shorthand for a tautological implication. The rules for con-


structing derivations are rules for inferring tautological implications from
one another, and a tautological implication can be correct regardless of
whether its premises are true. So adopting premises not known to be true
is not logically fallacious. It may just be pointless. And it will turn out
below that it need not even be pointless. There are legitimate reasoning
strategies that proceed by adopting premises not known to be true.
It will emerge that rules I and P must be supplemented by some
additional rules of inference, but we can do a fair amount of useful reasoning
using just I and P. A linear derivation is one constructed using just these
rules. For instance, suppose we want to show that (P ↔ Q) and (Q ↔ R)
jointly imply (P ↔ R). We could do this by constructing the following
derivation:

(1) 1. (P ↔ Q) premise
(2) 2. (Q ↔ R) premise
(1) 3. [(P → Q) & (Q → P)] (E7), 1
(2) 4. [(Q → R) & (R → Q)] (E7), 2
(1) 5. (P → Q) (I1), 3
(2) 6. (Q → R) (I1), 4
(1, 2) 7. (P → R) (I13), 5,6
(2) 8. (R → Q) (I2), 4
(1) 9. (Q → P) (I2), 3
(1, 2) 10. (R → P) (I13), 8,9
(1, 2) 11. [(P → R) & (R → P)] (I14), 7,10
(1, 2) 12. (P ↔ R) (E7), 11

In using rule I, the implications I1–I15 and the equivalences E1–E20


serve as “patterns”. The metalinguistic variables A, B, C , etc., are just
placeholders in the patterns, and can stand for any formulas at all. Thus in
the above derivation, in getting line 5 from line 4, we use I1:

(A & B) ¤ A.

For this purpose, we let A be the formula (P → Q) and we let B be the


formula (Q → P).
Typically, the purpose of constructing a derivation is to show that the
formula on the last line of the derivation follows from its premises. So a
derivation is said to be a derivation of that formula from those premises.
Thus, for example, the preceding example is a derivation of (P ↔ R) from
(P ↔ Q) and (Q ↔ R).
It should be emphasized that Rule I can only be applied to an entire
line. If, for example, (P & Q) is the formula on one line of a derivation, then
we can use Rule I with (I2) to obtain Q on a later line. But if [(P & Q) → R]
appears on a line of a derivation, we cannot use Rule I with (I2) to obtain
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 76

anything from it. In, particular, we cannot obtain (Q → R) by Rule I, using


(I2). In fact, (Q → R) is simply not implied by [(P & Q) → R].

Exercises

1. Construct a derivation of (P ∨ Q) from the premise (P & R).


2. Construct a derivation of (C & D) from the premises (P → C),
(Q → C), D, and (P ∨ Q).
3. Construct a derivation of (P ↔ Q) from the premises (~P → R), (R →
~Q), and (P → Q).

4. Suppositional Arguments
Although we can do a lot of reasoning using just rules I and P, many
of the implications we will want to establish cannot be proven using just
these rules. For most purposes, we need at least one additional rule. Recall
the principle of conditionalization, discussed in chapter three:
If C1 ,…,Cn, A ¤ B, then C 1,…,C n ¤ (A → B).
This tells us that we can remove one formula from the head of the implication
and make it the antecedent of a conditional in the conclusion of the implication.
There is a corresponding rule of inference that can be used in constructing
derivations in the propositional calculus:
RULE C: C ONDITIONALIZATION: If some formula, B appears on a line of a
derivation, and A is any premise of that line, then on any later line
of the derivation we can write the conditional (A → B), taking for
premise numbers all the premise numbers of B except for the line
number of A.
Rule C is essentially just a rule that allows us to say that if we have derived
a formula from a premise, then if that premise is true the formula is true.
The following is a simple illustration of the use of Rule C:

(1) 1. P premise
(1) 2. (P ∨ Q) (I3), 1
3. [P → (P ∨ Q)] C, 1, 2
From the premise P we derived the conclusion (P ∨ Q). Thus, if P is true,
then (P ∨ Q) is true. But this means that [P → (P ∨ Q)] is true; and this is
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 77

just what we get on line 3 using Rule C. Rule C allows us to eliminate a


premise by writing it explicitly as the antecedent of a conditional, thus
saying that if that premise is true, then the conclusion we derived from it is
true.
Arguments constructed using rule C are examples of suppositional
arguments. In suppositional reasoning, we make a supposition (adopt it as a
premise for the purpose of the reasoning), perform some reasoning that
depends upon the supposition, and then discharge the supposition by employing
a rule of inference that enables us to draw a conclusion that no longer
depends upon the supposition. Conditionalization is the simplest rule for
discharging a supposition. Given a conclusion that depends upon a
supposition, conditionalization permits us to infer a new conclusion (a
conditional) that no longer depends upon the supposition. Instead, the
supposition is built into the new conclusion as its antecedent. In section six
we will encounter another rule that allows us to discharge a supposition.
One thing that must be emphasized about the use of Rule C is that
you can only conditionalize a formula with respect to one of its premises. If
a formula does not occur among the premises of a given line, then you
cannot conditionalize that line with respect to that formula. For instance,
the use of Rule C on line 5 of the following purported derivation is incorrect:

(1) 1. (P & Q) premise


(1) 2. Q (I2), 1
(1) 3. (Q ∨ P) (I3), 2
(1) 4. P (I1), 1
5. [(Q ∨ P) → P] C, 3, 4

We cannot conditionalize in this way, because (Q ∨ P) is not one of the


premises of line 4.
An example of the correct use of Rule C is the following:

(1) 1. P premise
(2) 2. Q premise
(1,2) 3. (P & Q) (I14), 1 ,2
(1) 4. [Q → (P & Q)] C, 2, 3
5. {P → [Q → (P & Q)]} C, 1, 4

From the premises (i.e., suppositions) P and Q we derived the conclusion (P


& Q). So (P & Q) follows from premises (1) and (2). Then, eliminating
premise (2), we can say that—given premise (l)—if premise (2) is true then
(P & Q) is true; that is, [Q → (P & Q)] is true. So we can infer [Q → (P &
Q)] from the premise P. Then using Rule C again, we can say that if P is
true, then [Q → (P & Q)] is true; that is,
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 78

{P → [Q → (P & Q)]}
is true. This latter formula no longer depends upon any premises, because
we have taken the only premise of the consequent and written it explicitly
as the antecedent of the conditional.
In the preceding derivation, line 5 has no premise numbers. As indicated
earlier, this means it is a tautology. In any derivation, we must begin by
adopting a premise. There is no other way to get started. But the important
thing about suppositional reasoning is that it allows you to adopt a premise
“temporarily”, as a supposition, and then elminate dependence on it later
by discharging it. This is what makes it possible to construct derivations of
tautologies even though the reasoning must begin from premises.
In effect, there are two kinds of premises. There may be premises that
are given. They constitute the background knowledge we can use in our
reasoning. But other premises are not facts we know to be true but rather
suppositions that we adopt as premises for the purposes of the reasoning and
with the intent of discharging them later. The most common strategy used
in constructing a derivation of a conditional is to take the antecedent as a
premise, try to derive the consequent from it, and then conditionalize to
obtain the conditional and eliminate dependence on the premise. This
illustrates why it can be useful to adopt premises that are not known to be
true. They can instead be adopted for strategic reasons, provided we can
later eliminate dependence on them through conditionalization.
At any given point in a derivation, any formula can be adopted as a
premise, but most premises will not be useful. If a premise is not among
the premises we are initially given, then it will only be useful if it can later
be eliminated by conditionalization or some other discharge rule. And a
premise can only be eliminated by conditionalization if we have an interest
in obtaining a conditional having that premise as its antecedent. The
conditional might be the conclusion we are trying to derive. But it can also
happen that we are trying to derive a conclusion that is not a conditional,
but which can be derived from a conditional by using other inference rules.
For example, suppose we want to construct a derivation of the tautology (P
∨ ~(P & Q)) (i.e., a derivation from no premises). By (E12), (P ∨ ~(P & Q)) is
equivalent to (~(P & Q) ∨ P), and by (E4) the latter is equivalent to the
conditional ((P & Q) → P). The conditional ((P & Q) → P) can be obtained
using conditionalization. So the whole derivation is as follows:

(1) 1. (P & Q) premise


2. P (I1),1
3. ((P & Q) → P) C, 1,2
4. (~(P & Q) ∨ P) (E4), 3
5. (P ∨ ~(P & Q)) (E12), 4
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 79

Exercises

Part One.
1. Using conditionalization, construct a derivation (from no premises)
of the tautology [(P & Q) → (Q ∨ R)].
2. Construct a derivation of the formula (P ↔ ~~P).
3. Construct a derivation of the formula ~(P & ~P).

Part Two. In each of the following purported derivations, check the lines
for errors, and indicate the numbers of the lines on which errors occur. An
error is said to occur in a line if the line cannot be introduced in accordance
with the rules of inference. This means that in checking a line for errors,
you only look at that line to see whether it has the necessary relation to the
previous lines, and treat the previous lines as if they were all correct. If a
line is incorrect, try to find a way of correcting it (perhaps by adding additional
lines).

1. (1) 1. P premise
(2) 2. (P → Q) premise
(1, 2) 3. Q (I9), 1, 2

2. 1. (P ∨ Q) premise

3. (1) 1. (P ∨ Q) premise
(2) 2. ~Q premise
(1, 2) 3. P (I11), 1, 2
(1) 4. [ ~Q → (P ∨ Q)] C, 2, 1
(1, 2) 5. {P & [ ~Q → (P ∨ Q)]} (I14), 3, 4

4. (1) 1. ~(P & ~Q) premise


(1) 2. (P → Q) (E5), 1
(2) 3. (Q → R) premise
(1, 2) 4. (P → R) (I13), 2, 3

5. (1) 1. [(R ∨ S) ↔ ~T] premise


(1) 2. {[(R ∨ S) & ~T] ∨ [ ~(R ∨ S) & ~~T]} (E8), 1
(1) 3. {[ ~T & (R ∨ S)] ∨ [ ~~T & ~(R ∨ S)]} (Ell), 2
(1) 4. [ ~T ↔ (R ∨ S)] (E8), 3

6. (1) 1. [R ↔ ~(S ∨ ~T)] premise


DERIVATIONS IN THE PROPOSITIONAL CALCULUS 80

(2) 2. ~R premise
(1) 3. {[R → ~(S ∨ ~T)] & [ ~(S ∨ ~T ) → R]} (E7),1
(1) 4. [R → ~(S ∨ ~T)] (I1), 3
(1, 2) 5. ~~(S ∨ ~T) (I10), 2, 4
(1, 2) 6. (S ∨ ~T) (E1), 5
(7) 7. T premise
(1, 2) 8. (~T ∨ S) (E12), 6
(1, 2, 7) 9. S (I11), 7, 8
(1, 7) 10. (~R → S) C, 2, 9
(11) 11. (R ↔ S) premise
(11) 12. [(R → S) & (S → R)] (E7), 11
(11) 13. (S → R) (I2), 12
(1, 2, 7, 11) 14. R (I9), 9, 13
(1, 7, 11) 15. (R & ~R) (I14), 2, 14
(7, 11) 16. {[R ↔ ~(S ∨ ~T)] → (R & ~R)} C, 1, 15
(17) 17. R premise
18. (R → R) C, 17, 17
19. ~(R & ~R) (E5), 18
(7, 11) 20. ~[R ↔ ~(S ∨ ~T)] (I10), 16, 19
(11) 21. {T → ~[R ↔ ~(S ∨ ~T)]} C, 7, 20
(11) 22. (S → {T → ~[R ↔ ~(S ∨ ~T)]}) C, 9, 21
(11) 23. {(S & T) → ~[R ↔ ~(S ∨ ~T)]} (E15), 22
(11) 24. {[R ↔ ~(S ∨ ~T)] → ~(S & T)} (E16), 23

5. Strategies
Derivations provide an effective tool for investigating the validity of
arguments expressed in English. For example, consider the following argu-
ment:

1. If the river floods or heavy rains later in the summer, then our entire
wheat crop will be destroyed.
2. If our entire wheat crop will be destroyed then the community will be
bankrupt.
3. The river will flood if, and only if, there is an early thaw in the mountains.

4. Therefore, if there is an early thaw in the mountains, then our entire


wheat crop will be destroyed and the community will be bankrupt.

We can symbolize the form of this argument in the propositional calculus


DERIVATIONS IN THE PROPOSITIONAL CALCULUS 81

as follows:

1. [(F ∨ H) → D]
2. (D → B)
3. (F ↔ E)

4. [E → (D & B)]

We could evaluate this argument by constructing its corresponding


conditional and then writing out a truth table, but such a truth table would
be 32 lines long and would have 352 entries. It is much easier to construct a
derivation of the conclusion from the premises as follows:

(1) 1. [(F ∨ H) → D] premise


(2) 2. (D → B) premise
(3) 3. (F ↔ E) premise
(4) 4. E premise
(3) 5. [(F → E) & (E → F)] (E7), 3
(3) 6. (E → F) (I2), 5
(3, 4) 7. F (I9), 4, 6
(3, 4) 8. (F ∨ H) (I3), 7
(1, 3, 4) 9. D (I9), 1, 8
(1, 2, 3, 4) 10. B (I9), 2, 9
(1, 2, 3, 4) 11. (D & B) (I14), 9, 10
(1, 2, 3) 12. [E → (D & B)] C, 4, 11

“This is all very well,” you may reply, “but how does one find
derivations?” The answer is that it requires ingenuity. The construction of
derivations is not so mechanical as the construction of truth tables. However,
there are definite strategies that are helpful in constructing derivations. If
these strategies are employed systematically, it is always possible to construct
a derivation of a formula from a set of premises that tautologically implies
it. We can list strategies for each type of formula (for example, conjunction,
disjunction, and so on).

5.1 Conditionals
We might be confronted with either of two situations with regard to
conditionals. We might be trying to get a conditional as a consequence of
something else, or we might be trying to get something else as a consequence
of a conditional. These cases are treated separately.

Backward reasoning
Let us suppose first that we are trying to derive a conditional from
something else. The last derivation above illustrates what is generally the
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 82

simplest way to go about this. In deriving the conditional [E → (D & B)]


from the premises, what we did was take the antecedent E as a new premise,
and then with its help we derived the consequent (D & B). Then we
conditionalized to get the conditional. This is almost always the best way
to derive a conditional from its premises. Let us look at another example of
this same technique. Suppose we want to show that
{Q → [R → (Q & R)]}
is a tautology. Then we can construct a derivation as follows:

(1) 1. Q premise
(2) 2. R premise
(1, 2) 3. (Q & R) (I14), 1, 2
(1) 4. [R → (Q & R)] C, 2, 3
5. {Q → [R → (Q & R)]} C, 1, 4

In constructing this derivation, we first took the antecedent of the conditional,


P, as a premise, and then sought to derive the consequent [R → (Q & R)]
from it. But this consequent is also a conditional, so we took its antecedent,
R, as a new premise and then sought to derive its consequent (Q & R). This
we could do using I14. Then we conditionalized in reverse order to attach
the antecedents in their proper places.
The derivation is an explanation of why the conclusion follows from
the premises, but it does not fully explain the reasoning that led us to the
derivation. That reasoning was bidirectional. Some of it proceeded backward
from the conclusion that we were trying to get to other formulas from
which we could get the conclusion. These other formulas became interests.
For example, because we were trying to get the conditional {Q → [R → (Q
& R)]}, we adopted Q as a supposition and then adopted an interest in
getting the consequent [R → (Q & R)]. Then we reasoned backward again
from that interest to an interest in getting (Q & R) given the combined
suppositions Q and R. At that point we were able to reason forward to (Q &
R). Then we completed the derivation by reversing the earlier backward
reasoning to draw the conclusions that we were looking for at each step of
the backward reasoning. When we draw conclusions by reversing the
backward reasoning, we say that we are discharging the interests. The interests
were for something, and we are now getting it.

Annotated derivations
A derivation only documents the forward reasoning and the interest
discharges in our reasoning. To help keep track of what we have done in
searching for a derivation, it is often useful to include more information by
indicating how the backward reasoning went as well. For this purpose we
can write annotated derivations. For example, we can annotate the preceding
derivation as follows:
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 83

1. {Q → [R → (Q & R)]}
(1) 1. Q premise
(1) 2. [R → (Q & R)] for interest 1 by C using premise (1)
(2) 2. R premise
(1,2) 3. (Q & R) for interest 2 by C using premise (2)
(1,2) 4. Q for interest 3 by (I14) discharged by (1)
(1,2) 5. R for interest 3 by (I14) discharged by (2)
(1, 2) 3. (Q & R) (I14), 1, 2 discharges interest 3
(1) 4. [R → (Q & R)] C, 2, 3 discharges interest 2

5. {Q → [R → (Q & R)]} C, 1, 4 discharges interest 1

The bold italicized text records the construction and discharge of interests.
The interests are numbered sequentially and offset to the right to distinguish
them from conclusions. To the right of each interest is an explanation of
why it was adopted. In that explanation, numbers in parentheses refer to
conclusions and numbers not in parentheses refer to earlier interests. The
list of premise numbers preceding an interest indicates which premises can
be used in deriving a conclusion that discharges the interest. Interest 2 is an
interest in obtaining the consequent of the conditional in interest 1, so we
can also use the antecedent of the conditional (premise 1) in getting the
consequent for the purpose of employing conditionalization. Similarly for
interest 3.
With one exception, when an interest is added to an annotated derivation
by reasoning backward from another interest, the new interest has the same
premise numbers as the interest from which it was obtained. The exception
concerns interests adopted by conditionalization. Such an interest is
accompanied by a new premise (the supposition of the antecedent of the
desired conditional), and the number of that new premise is added to the
premise numbers of the interest in the consequent of the conditional. This
is illustrated in interests 2 and 3 of the preceding annotated derivation.
This represents the fact that for purposes of conditionalization, we can use
the premise recording the supposition of the antecedent in trying to derive
the consequent. Rule R, which will be introduced in section seven, works
similarly. That is, it introduces a new premise (a supposition) which can
then be used in trying to derive the interests generated by rule R.
The bidirectional reasoning recorded in an annotated derivation is
called interest driven reasoning because the structure of the reasoning is driven
as much by what we are trying to prove as by whatever premises we might
be given.

Forward reasoning
When we are trying to derive a conditional from something else, we
are engaged in backward reasoning and subsequent interest discharge. If
we are instead trying to derive something from a conditional, we are engaged
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 84

in forward reasoning. In the case of conditionals, there are four common


varieties of forward inferences. The simplest and most common forward
inference using conditionals is modus ponens. If we have a conditional (A →
B) on one line and we have the antecedent A on another line, then by I9 we
can derive B. An example would be the following:

(1) 1. (P & Q) premise


(2) 2. [Q → (R & ~S)] premise
(1) 3. Q (I2), 1
(1, 2) 4. (R & ~S) (I9), 2, 3
(1, 2) 5. ~S (I2), 4

Another common case is that in which we have a conditional (A → B)


on one line, and we have ~B, the negation of its consequent, on another
line. Using I10 we can derive ~A, the negation of the antecedent. This
inference is called modus tollens. The following contains an example of
modus tollens:

(1) 1. [Q → (R & S)] premise


(2) 2. ~S premise
(2) 3. (~R ∨ ~S) (I4), 2
(2) 4. ~(R & S) (E2), 3
(1, 2) 5. ~Q (I10), 1, 4

A less common forward inference occurs when we have a conditional


of the form (A → B) and have some other conditional of the form ( B → C).
From these, using I13, we can derive (A → C). This inference is called
hypothetical syllogism. An example of the use of hypothetical syllogism is
found in the following derivation:

(1) 1. (~P ∨ Q) premise


(2) 2. (~Q ∨ R) premise
(1) 3. (P → Q) (E4), 1
(2) 4. (Q → R) (E4), 2
(1, 2) 5. (P → R) (I13), 3, 4
(1, 2) 6. (~P ∨ R) (E4), 5

Another common inference from conditionals takes one of the following


forms:
(A → B) ¤ (~B → ~A)
(~A → ~B) ¤ (B → A)
These inferences are both called contraposition, and are licensed by E16.
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 85

Exercises for section 5.1

For each of the following, give an annotated derivation of the formula


below the line from the formulas above the line. Keep in mind the strategies
for constructing derivations.

1. [P → (Q → R)] 2. (P → Q)
[(P → Q) → (P → R)] (~P → R)
(~R → Q)

5.2 Conjunctions
The strategies for reasoning forward or backward with conjunctions
are simple and basically similar. When reasoning forward from a conjunction
(A & B), it is generally a good idea to “take the conjunction apart”, using I1
and I2 to obtain the individual conjuncts and then reason from them.
Conversely, when trying to infer a conjunction from other formulas, it is
generally best to try to obtain the conjuncts separately and then conjoin
them using I14. Both of these strategies are illustrated by the following
annotated derivation of [(Q ∨ R) & (P ∨ S)] from (P & Q):

(1) 1. (P & Q) premise


(1) 1. [(Q ∨ R) & (P ∨ S)]
(1) 2. (Q ∨ R) 
 for interest 1 by (I14).
(1) 3. (P ∨ S) 
(1) 2. P (I1), 1
(1) 3. Q (I2), 1
(1) 4. (P ∨ S) (I3), 2 this discharges interest 3
(1) 5. (Q ∨ R) (I3), 3 this discharges interest 2
(1) 6. [(Q ∨ R) & (P ∨ S)] (I14),4,5 this discharges interest
1

Exercise for section 5.2

Give an annotated derivation of the formula below the line from the formulas
above the line. Keep in mind the strategies for constructing derivations.

1. (P → Q)
(P → R)
[P → (Q & R)]
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 86

5.3 Disjunctions
There is a variety of strategies for use with disjunctions. The strategies
for backward reasoning are different from those for forward reasoning, so
we will discuss them separately.

Backward reasoning
The strategies for backward reasoning are strategies for deriving
disjunctions from other formulas. The simplest case occurs when we can
derive one of the disjuncts separately and then infer the disjunction by I3 or
I4 (addition). This was illustrated by the preceding derivation. We could
expand its annotation as follows:

(1) 1. (P & Q) premise


(1) 1. [(Q ∨ R) & (P ∨ S)]
(1) 2. (Q ∨ R) 
 for interest 1 by (I14)
(1) 3. (P ∨ S) 
(1) 4. Q for interest 2 by (I3)
(1) 5. P for interest 3 by (I3)
(1) 2. P (I1), 1 this discharges interest 5
(1) 3. Q (I2), 1 this discharges interest 4
(1) 4. (P ∨ S) (I3), 2 this discharges interest 3
(1) 5. (Q ∨ R) (I3), 3 this discharges interest 2
(1) 6. [(Q ∨ R) & (P ∨ S)] (I14),4,5 this discharges interest 1

It is rather uncommon to be able to derive a disjunction using addition.


A strategy that is usually more effective employs either E19 or E20. These
equivalences tell us that a disjunction is equivalent to a conditional. We
observe that (A ∨ B) is equivalent to (~ A → B) by E19 and it is equivalent
to (~B → A) by E20. So to get the disjunction, we adopt interest in one of
these conditionals. For this purpose, we can employ the strategy for
conditionals to get the conditional. Thus to obtain (A ∨ B) we can take the
negation of one of the disjuncts as a premise (as a supposition), use that
premise to derive the other disjunct, conditionalize to obtain a conditional,
and then convert the conditional back into a disjunction using E19 or E20.
To illustrate, suppose we want to derive (P ∨ Q) from [P ∨ (Q & R)]. We
could proceed as follows:

(1) 1. [P ∨ (Q & R)] premise


(1) 1. (P ∨ Q)
(1) 2. (~P → Q) E19, for interest 1
(2) 2. ~P premise
(1,2) 3. Q for interest 2 by C using premise (2)
(1, 2) 3. (Q & R) (I11), 1, 2
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 87

(1, 2) 4. Q (I1), 3 this discharges interest 3


(1) 5. (~P → Q) C, 2, 4 this discharges interest 2
(1) 6. (P ∨ Q) E19, 5 this discharges interest 1

Forward reasoning
In reasoning forwards from a disjunction, one strategy is to somehow
obtain the negation of one of the disjuncts and then infer the other disjunct
in accordance with I11 or I12 (disjunctive syllogism). This was the strategy
employed on line 3 of the preceding derivation. However, one is not usually
so fortunate as to be able to derive the negation of one of the disjuncts, so
this strategy is not often applicable.
A strategy that can be used in general for reasoning forward from a
disjunction proceeds in terms of I15 (dilemma). That is, if we want to derive
some conclusion C from a disjunction of the form (A ∨ B), we can do that
by deriving the two conditionals (A → C) and (B → C ). In other words,
we derive our desired conclusion C separately from each disjunct, and then
use I15 to get C . An example of this would be the following:

(1) 1. [(P & Q) ∨ (P & R)] premise


(1) 1. P
(1) 2. [(P & Q) → P]  for interest 1 by (I15)

(1) 3. [(P & R) → P]  using premise (1)
(2) 2. (P & Q) premise
(1,2) 4. P for interest 2 by C using premise (2)
(2) 3. P (I1), 2 this discharges interest 4
4. [(P & Q) → P] C, 2,3 this discharges interest 3
(5) 5. (P & R) premise
(1,5) 5. P for interest 3 by C using premise (5)
(5) 6. P (I1), 5 this discharges interest 5
7. [(P & R) → P] C,5,6 this discharges interest 3
(1) 8. P (I15), 1, 4, 7 this discharges interest 1

In this annotated derivation notice that we have two separate interests in P.


This is because in inferring a conclusion that discharges interest 4 we are
allowed to use supposition 2, whereas in inferring a conclusion that discharges
interest 5 we are not allowed to use supposition 2 but we can use supposition
5 instead.
The use of dilemma to derive conclusions from disjunctions is usually
effective, but rather inefficient. The problem with this strategy is that, in
effect, one must construct two separate derivations of the conclusion, one
from each disjunct. So this should be regarded as a kind of “last resort”
strategy. Try this only if you cannot find a better way to do the derivation.
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 88

Exercise for section 5.3

Give an annotated derivation of the formula below the line from the formula
above the line. Do this in two different ways: (1) Use the strategy for
deriving a disjunction from something else, and disjunctive syllogism; (2)
use dilemma:

1. [Q ∨ (P & R)]
(Q ∨ R)

5.4 Biconditionals
The strategy for dealing with biconditionals is basically simple—turn
them into somethng else that we already know how to deal with. We have
two equivalences that are useful for this purpose, E7 and E8. By E7 the
biconditional (A ↔ B) is equivalent to the conjunction [(A → B) &
(B → A)], so if we wish to derive something from a biconditional, we can
first transform the biconditional into a conjunction, using E7, and then proceed
as with conjunctions. Conversely, if we wish to derive a biconditional from
something else, we can first derive the conjunction to which it is equivalent,
and then use E7 to transform it into the biconditional. The conjuncts are
conditionals, so we will typically use conditionalization to derive them.
Here is an example, in which we derive (~P ↔ ~Q) from (P ↔ Q):

(1) 1. (P ↔ Q) premise
(1) 1. (~P ↔ ~Q)
(1) 2. [(P → Q) & (Q → P)] E7, 1
(1) 3. (P → Q) (I1), 2
(1) 4. (Q → P) (I2), 2
(1) 2. [(~P → ~Q) & (~Q → ~P)] E17, for interest 1
(1) 3. (~P → ~Q) 
 for interest 2 by (I14)
(1) 4. (~Q → ~P) 
(1) 5. (~P → ~Q) E16, 4 this discharges interest 3
(1) 6. (~Q → ~P) E16, 3 this discharges interest 4
(1) 7. [(~P → ~Q) & (~Q → ~P)] (I14), 5, 6
this discharges interest 2
(1) 8. (~P ↔ ~Q) E7, 7 this discharges interest 1

Alternatively, we can use E8 to transform a biconditional (A ↔ B) into a


disjunction [(A & B) ∨ (~A & ~B)], and then work with the disjunction.
This is sometimes effective, but usually less useful than the preceding strategy.
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 89

Exercises for section 5.4

For each of the following, give an annotated derivation of the formula


below the line from the formulas above the line. Keep in mind the strategies
for constructing derivations.

1. (A ↔ B) 2. (A ↔ B)
~(A & B) (B ↔ C)
~B (A → C)

5.5 Negations
A general strategy for dealing with negations is to use one of the
equivalences E1, E2, E3, E5, or E6 to convert the negated formula into
something else. This works for the negation of anything other than an
atomic formula. The negation of a negation drops both negations, the negation
of a disjunction becomes a conjunction, the negation of a conjunction becomes
a disjunction, the negation of a conditional becomes a conjunction, and the
negation of a biconditional becomes another biconditional. Thus we can
eliminate the negation and use the strategies that are applicable to the
equivalent formula to which the negation is transformed. For example, if
we want to derive (~P ∨ ~Q) from ~(P ∨ Q), we might proceed as follows:

(1) 1. ~(P ∨ Q) premise


(1) 1. (~P ∨ ~Q)
(1) 2. ~P for interest 1 by (I3)
(1) 2. (~P & ~Q) (E2), 1
(1) 3. ~P (I1), 2 this discharges interest 2
(1) 4. (~P ∨ ~Q) (I3), 3 this discharges interest 1

Exercises for section 5.5

For each of the following, give an annotated derivation of the formula


below the line from the formulas above the line. Keep in mind the strategies
for constructing derivations.

1. ~[P → (Q & R)] 2. [P & ~(R ∨ S)]


R ~(P → S)
(P → ~Q)

5.6 Combining the Strategies


DERIVATIONS IN THE PROPOSITIONAL CALCULUS 90

Let us consider an extended example illustrating how these strategies


can be combined in complex derivations that use a number of strategies
simultaneously. Suppose we want to derive (S → ~T) from the premise
{~[P → (S & T )] ∨ [(S → ~P) & (T → P)]}. We can do this with the
following derivation:

(1) 1. { ~[P → (S & T) ] ∨ [(S → ~P) & (T → P)]} premise


(1) 1. (S → ~T)

(2) 2. {~[P → (S & T)] → (S → ~T)}  for interest



 1 by (I15)

(3) 3. {[(S → ~P) & (T → P)] → (S → ~T)}  using (1)
(2) 2. ~[P → (S & T)] premise
(4) 4. (S → ~T) for interest 2 by C using premise (2)
(2) 3. [P & ~(S & T)] (E5), 2
(2) 4. ~(S & T) (I2), 3
(2) 5. (~S ∨ ~T) (E3), 4
(2) 6. (S → ~T) (E4), 5 this discharges interest 4
7. {~[P → (S & T) ] → (S → ~T)} C, 2, 6 this discharges interest 2
(8) 8. [(S → ~P) & (T → P)] premise
(8) 5. (S → ~T) for interest 3 by C using premise (8)
(9) 9. S premise
(8,9) 6. ~T for interest 5 by C using premise (9)
(8,9) 10. (S → ~P) (I1), 9
(8,9) 11. ~P (I9), 9, 10
(8,9) 12. (T → P) (I2), 8
(8,9) 13. ~T (I10), 11, 12 this discharges interest 6
(8) 14. (S → ~T) C, 9, 13 this discharges interest 5
15. {[(S → ~P) & (T → P)] → (S → ~T)} C, 8, 14
this discharges interest 3
(1) 15. (S → ~T) (I15), 1, 7, 15 this discharges interest 1

Here we have followed the dilemma strategy for deriving a conclusion


from a disjunction. The premise is a disjunction, so we tried to derive the
conclusion from each disjunct separately, which we did on lines 7 and 15,
and then used I15 to obtain the conclusion from our initial premise. We
were trying to obtain conditionals on lines 7 and 15, so we used the strategy
for deriving conditionals, which is to take the antecedent as a new premise,
and then try to derive the consequent from it. Thus on line 2 we took the
first disjunct as a new premise. This disjunct is a negation, so we used our
strategy for negations, which is to turn them into something else. We did
this using E5. The result is a conjunction. From a conjunction we can
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 91

obtain each conjunct separately. In this case we only needed the second
conjunct, which we obtained by I2: ~(S & T). This is a negation, so we used
E3 to drive the negation in and convert it to a disjunction (~S ∨ ~T). We
were looking for (S → ~T), which we could then obtain by (E4). Thus we
derived (S → ~T) from the first disjunct. Next we turned to the second
disjunct, taking it as a new premise. It is a conjunction, so we obtained each
conjunct separately, using I1 and I2. To obtain (S → ~T) we took the
antecedent as a premise and tried to derive the consequent. That was done
immediately by I9 and I10. Thus we obtained (S → ~T) from each disjunct
of line 1, and hence by I15 it followed from line 1.

Exercises

For each of the following, give an annotated derivation of the formula


below the line from the formulas above the line. Keep in mind the strategies
for constructing derivations.

1. [P → (Q → R)] 2. [P → (~Q → ~P)]


~R ~(P → Q)
(P → ~Q) ~P

3. ~(R → S) 4. (~P → R)
[R → (S ∨ ~T)] [ ~Q → (R → S)]
~T [(P ∨ Q) ∨ (R & S)]

5. P 6. [(R → S ) ∨ (R & ~T)]


[(P & Q) ∨ (P & ~Q)] [R & (T ↔ ~S)]
~T

7. (P & ~P) 8. ~[R ∨ (~S → ~T)]


Q ~[(R → S) → ~(~R → ~Q)]
~Q

6. Double Negations
The rules and strategies discussed above are adequate for deriving
formulas from any premises that tautologically imply them. However, the
derivations produced in this way are often more complex than they need to
be. To simplify things, we will add two more rules of inference. The rule
DN of double negation will be presented in this section, and reductio ad
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 92

absurdum will be discussed in the next section.


Many inferences produce formulas containing double negations, and
for the reasoning to proceed we usually have to eliminate the double
negations. That can be tedious using the rules presented above. To illustrate,
suppose we want to show that (~P → P) tautologically implies P. To construct
a derivation, we might proceed as follows:

(1) 1. (~P → P) premise


1. P
(1) 2. (~~P ∨ P) (E4), 1
(1) 2. (~~P → P)  for interest 1 by (I15)

(1) 3. (P → P)  using conclusion (2)
(4) 4. ~~P premise
(1,4) 4. P for interest 2 by C using premise (4)
(4) 5. P (E1), 4 this discharges interest 4
6. (~~P → P) C, 4, 5 this discharges interest 2
(7) 7. P premise
(1,4) 5. P for interest 3 by C using premise (7)
discharged by (7)
8. (P → P) C, 7, 7
(1) 9. P (I15), 1, 6, 8 this discharges interest 1

This reasoning is made complex by the fact that to eliminate the double
negation in (~~P ∨ P), we must get ~~P on a line by itself and then use (E1).
If we could eliminate the double negation inside the disjunction, we could
reason as follows:

(1) 1. (~P → P) premise


1. P
(1) 2. (~~P ∨ P) (E4), 1
(1) 3. (P ∨ P) eliminating double negation in 2
(1) 4. P (E17), 3 this discharges interest 1

The inference from (2) to (3) is not licensed by rule I, beause as we saw
above, rule I can only be applied to the entire formula on a line—not to
parts of a formula. However, in this case the inference is clearly valid. The
following metatheorem holds in general:
If A is a formula containing a double negation ~~B as one of its
parts, and A* results from replacing the occurrence of ~~B by B,
then A is tautologically equivalent to A*.
It is easy to see that this principle holds. Just think about the truth tables
for A and A*. Because ~~B and B are equivalent, they will have the same
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 93

truth value on any given line of the truth table for A or A*. The other parts
of the formulas are identical, so they will also have the same truth values,
and hence the truth values for A and A* must come out the same on each
line. So A and A* are tautologically equivalent.
This metatheorem justifies the adoption of an additional inference rule
allowing us to introduce or eliminate double negations as we see fit:
RULE DN: D OUBLE NEGATION:
If some formula, A, appears on a line of a derivation, and a double
negation ~~B is one of its parts, then where A* results from replacing
a single occurrence of ~~B by B, A* can be written on any later line
of the derivation, taking for premise numbers all the premise numbers
of A.
If some formula, A, appears on a line of a derivation, and a formula
B is one of its parts, then where A* results from replacing a single
occurrence of B by ~~B, A* can be written on any later line of the
derivation, taking for premise numbers all the premise numbers of
A.
Notice that DN has two parts. The first part allows us to eliminate a double
negation whenever we want, and the second part allows us to introduce a
double negation whenever that it strategically desirable. Notice also that
DN only allows us to introduce or eliminate a single double negation at a
time. To introduce or eliminate several double negations requires several
applications of the rule.

Exercises

For each of the following, give two derivations of the formula below the
line from the formulas above the line. Keep in mind the strategies for
constructing derivations. In the first derivation, use rule DN. In the second
derivation, do not use rule DN.

1. (P → ~~Q) 2. (P → ~~Q)
(~Q → ~P) (Q → ~~R)
(P → ~~R)

7. Reductio ad Absurdum
A very useful strategy that we have not yet discussed is called reductio
ad absurdum (reduction to absurdity). This is actually a pair of related
strategies, both based upon the same idea. The idea is that to show that
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 94

something is true, it suffices to show that the supposition that it is false


leads to an absurd result. This can be turned into a precise strategy by
noting the following two tautological implications:
(~P → P) ¤ P
[~P → (Q & ~ Q)] ¤ P
At the beginning of the preceding section we showed how to derive P from
(~P → P). P can be derived from [~P → (Q & ~Q)] as follows:

(1) 1. [~P → (Q & ~Q)] premise


(1) 1. P
(1) 2. ~(Q & ~Q) (I10), for interest 1 by (I10)
(1) 3. (~Q ∨ ~~Q) for interest 2 by (E3)
(1) 4. (Q → ~~Q) for interest 3 by (E4)
(2) 2. Q premise
(2) 5. ~~Q for interest 4 by C using premise (2)
(2) 6. Q for interest 5 by (E1) discharged by (2)
(2) 3. ~~Q (E1), 2 this discharges interest 5
4. (Q → ~~Q) C, 2, 3 this discharges interest 4
5. (~Q ∨ ~~Q) (E4), 4 this discharges interest 3
6. ~(Q & ~Q) (E3), 5 this discharges interest 2
(1) 7. ~~P (I10), 1, 6
(1) 8. P (E1), 7 (or by DN, 7) this discharges interest 1

The strategy of reductio ad absurdum then consists of deriving either (~P →


P) or a formula of the form [~P → (Q & ~Q)], in order to derive P. The
derivation for (~P → P) can then be turned into a derivation of P by
adding the derivation of P from (~P → P) to the end of it. And the
derivation for [~P → (Q & ~Q )] can be turned into a derivation of P by
adding the derivation of P from [~P → (Q & ~Q )] to the end of it. Note
that a slight variant of the strategy will allow us to derive ~P by deriving
(P → ~P) or by deriving [P → (Q & ~ Q)].
This strategy is extremely useful, but it is tedious to have to finish the
derivations by repeatedly rewriting the derivation of P from either (~P →
P) or [~P → ( Q & ~Q )]. We could rectify that by adding the two tautological
implications:
(~P → P) ¤ P
[~P → (Q & ~ Q)] ¤ P
to the list of implications used by rule I. That would allow us to infer P
directly from either (~P → P) or [~P → (Q & ~Q )]. However, we can
simplify things even more by adding a rule of inference (really, a pair of
rules) instead:
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 95

RULE R: REDUCTIO AD ABSURDUM:


If a formula Q appears on some line of a derivation and ~Q appears
on another line, then for any formula P that is a premise of one of
those lines, ~P can be written on any later line of the derivation,
taking for premise numbers all of the premise numbers of Q and ~Q
except for the line number of P.
If a formula Q appears on some line of a derivation and ~Q appears
on another line, then for any formula ~P that is a premise of one of
those lines, P can be written on any later line of the derivation,
taking for premise numbers all of the premise numbers of Q and ~Q
except for the line number of ~P.
The first version of the rule is of use when we are trying to derive a negation.
Otherwise fwe use the second version. By using this rule we avoid having
to conjoin Q and ~Q. and we also avoid having to conditionalize to construct
the conditional. Notice also that this rule accommodates both kinds of
reductio ad absurdum. If we infer P from the supposition ~ P, then we
automatically have a formula and its negation appearing on separate lines
of the derivation, and so we can apply the rule. This is just the case in
which Q is P itself.
Consider an example of the use of this rule. Suppose we want to
derive [(P → Q) → (P → R)] from the premise [P → (Q → R)]. We could
do it as follows:

(1) 1. [P → (Q → R)] premise


(1) 1. [(P → Q) → (P → R)]
(2) 2. (P → Q) premise
(1,2) 2. (P → R) for interest 1 by C using premise (2)
(3) 3. ~(P → R) premise (for rule R)
(3) 4. (P & ~R) (E5), 3
(3) 5. P (I1), 4
(3) 6. ~R (I2), 4
(1,3) 7. (Q → R) (I9), 1, 5
(2,3) 8. Q (I9), 2, 5
(1,2,3) 9. R (I9), 7, 8
(1,2) 10. (P → R) R, 6, 9 this discharges interest 2
(1) 11. [(P → Q) → (P → R)] C, 2, 10 this discharges interest 1

This illustrates the use of rule R, but the astute reader may notice that it
would have been easier to get the same conclusion by using the strategy for
conditionals, taking P as a premise and deriving R. Typically, derivations
based upon reductio ad absurdum will be long, and if there is another way to
solve the problem it is better to do it that way. Like the use of dilemma,
reductio ad absurdum should be viewed as a “last resort” strategy, to be used
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 96

only when simpler strategies fail.


Notice that rule C, rule R is a discharge rule. That is, it allows us to
draw a conclusion having one fewer premise than the lines from which the
conclusion is drawn. This is because rule R is actually a shortcut for a
pattern of reasoning that includes a use of rule C. Doing the same reasoning
without rule R would require us to derive the conditional [~P → ( Q &
~Q)], which we would normally do by conditionalization.
What reasoning in accordance with rule R actually amounts to is a
demonstration that a set of premises is inconsistent. If we add ~P to our
given premises and then derive both Q and ~Q from that enlarged set of
premises, we have shown that those premises cannot all be true. Thus the
negation of any one of them follows deductively from the others. In particular,
P follows from the given premises.

Exercises

Part (a) For each of the following, give an annotated derivation of the formula
below the line from the formulas above the line. Keep in mind the strategies
for constructing derivations. Try doing the derivations both with and without
rule R.

1. (P → Q) 2. [P → (Q → R)]
~(P → R) (P → Q)
~(Q → R) (R → ~Q)
~P

3. [P → ~(S & ~T)]


(~P → ~S)
~T
~S

Part (b) For each of the following, give an annotated derivation of the
formula below the line from the formulas above the line.

1. (S → ~S) 2. ~{P → [ ~Q & ~(R ∨ ~S)]}


(R ∨ S) {[P & (R ∨ ~S )] ∨ (P & Q)}
(P → ~R)
[(P → S) & ~S]

3. [(P ∨ Q) ∨ R] 4. [(P ∨ Q) → (P ∨ R)]


[P → (S ↔ ~T)] [P ∨ (Q → R)]
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 97

[Q → ~(S ∨ ~T)]
[T → ~(R & S)]
[ ~T → (R & S)]
(S ↔ ~T)

Part (c) Show that the following formulas are tautologies by giving annotated
derivations of them; that is, derivations the last lines of which have no
premise numbers.
1. (P ∨ ~P)
2. [(P & Q) → (~Q → R)]
3. [(P ↔ Q ) → (~P ↔ ~Q)]
4. {(P → Q) → [(Q → R ) → (P → R)]}
5. [P → (~P → Q)]
6. [P → (Q → P)]
7. {[(P & Q) → R] ↔ [(P & ~R) → ~Q]}
8. {P ↔ [~P → (Q & ~Q)]}
9. [(P → Q) ∨ (Q → P)]

Part (d) For each of the following, if the formula below the line is tautologically
implied by the formulas above the line, construct a derivation to show that
it is. If the formula below the line is not tautologically implied by the
formulas above the line, show that it is not by a suitable assignment of truth
values to the atomic parts of the formulas.

1. [P → (~Q → R)] 2. [P ↔ (R ∨ ~P)]


~R R
(P → Q)

3. [P → (~Q → R)] 4. [P ↔ (R ∨ ~P)]


Q ~R
~(P → R)

Part (e) Symbolize each of the following arguments and construct a derivation
of the conclusion from the premises:

1. If John comes to the party, Joe will not come. Joe will not come unless
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 98

Mary comes, but Mary will only come if John comes. Therefore, Joe
will not come to the party.

2. If the bank rejects his loan application, Mr. Horner will either have to
get the money somewhere else or sell his business. If the bank will not
lend him the money, he will not be able to get it anywhere else either.
But if he sells his business, his only alternative will be to join the Peace
Corps. Thus if the bank rejects his loan application, Mr. Horner will
join the Peace Corps.

3. If the Stanford beats UCLA in basketball, they will score a great victory.
If Stanford announces that they are cutting their budget for sports,
UCLA will cut its sports budget. If UCLA cuts its sports budget, but
the Stanford does not cut theirs, then Stanford will beat UCLA in
basketball. Stanford is very devious, and will announce that they have
cut their budget without actually doing so. Therefore, Stanford will
score a great victory.

4. If the class does well, and the professor knows it, he will not give them
a test. But the professor will not know that the class does well unless
he gives them a test. Therefore, if the class does well the professor will
not know it.

5. If Didereux becomes President of France, England will lose its influence


in the European Union. If England loses its influence in the European
Union, she will have to trade heavily with the United States to avoid
economic disaster. If England trades heavily with the United States
and loses its influence in the European Union, many large corporations
will move out of England. If many large corporations move out of
England, she will suffer economic disaster. If England suffers economic
disaster, she will be taken over by the Communists. Consequently, if
Didereux becomes President of France, England will be taken over by
the Communists.

Part (f) Try your hand at the following derivations, but don’t be disappointed
if you do not get them. They are difficult.

1. (Q → R)
[R → (P & Q)]
[P → (Q ∨ R)]
(P ↔ Q)

2. no premises
[{(P ∨ Q) & [(~P ∨ Q) & (P ∨ ~Q)]} → ~(~P ∨ ~Q)]

3. no premises
{[(P ↔ Q) ↔ R] ↔ [P ↔ (Q ↔ R)]}
DERIVATIONS IN THE PROPOSITIONAL CALCULUS 99

4. no premises
({[P & (Q → R)] →S} ↔ {[~P ∨ (Q ∨ S)] & [~P ∨ (~R ∨ S)]})
Part Two:

The
Predicate
Calculus
5
The Syntax of the
Predicate Calculus

1. Introduction to the Predicate Calculus


We have seen that we can deal with many statement forms and argument
forms within the propositional calculus. But there are also many statement
forms and argument forms that do not fall within the scope of the
propositional calculus. For instance, we have no way of expressing the
form of statements like “All men are mortal” in the propositional calculus.
This statement has the form “All A are B”, but that cannot be symbolized as
a formula of the propositional calculus. And similarly, the argument form

1. All A are B.
2. All B are C.

3. Therefore, all A are C.

is valid, but cannot be expressed in the propositional calculus. In general,


those statement forms and argument forms that involve such words as
“all”, “each”, “every”, and “some” or “there is”, cannot be symbolized in
the propositional calculus. To study these statement and argument forms,
we must extend the propositional calculus, making it more expressive. The
logical theory that we get when we do this is called the predicate calculus. It
turns out that virtually all of the deductive reasoning we perform can be
expressed in the predicate calculus.

2. Predicates, Relations,
and Individual Constants
To deal with the forms of statements like “All men are mortal” or “All
apples are red”, we need the concept of a predicate. When we say something
about some individual (a person, object, number, and so on), we ascribe a
predicate to that individual. If we say “John is a bachelor”, we ascribe the
predicate “is a bachelor” to John. In order to symbolize the form of a
statement like “John is a bachelor”, we use capital letters, with or without
numerical subscripts, to stand for predicates. We might use B to stand for
“is a bachelor”. We must also have some way of symbolizing those
expressions, like “John”, or “the girl with the red hair”, that denote
individuals. For this purpose we use the lower case letters from a through
s, with or without numerical subscripts. They are called individual constants.
We might let j stand for “John”. Using predicate letters and individual
constants we can symbolize the form of the statement “John is a bachelor”
THE SYNTAX OF THE PREDICATE CALCULUS 103

as Bj. In other words, we write the predicate letter followed immediately


by the individual constant. Whenever we have a statement of the form “b
is A” (for example, “This apple is red”, or “James is angry”), we symbolize
it in the predicate calculus at Ab.
Predicates are used to say something about a single individual. But
sometimes we want to talk about two or more individuals, saying how they
are related. For this purpose we use relations. For example, the statement
“John is taller than Bill” employs the relation “is taller than”. We might
symbolize this as Tjb, letting j stand for “John”, b for “Bill”, and T for the
relation “is taller than”. “is taller than” is a two-place relation, i.e., it relates
two things. There can also be three-place relations, four-place relations, etc.
For example, “is between” relates three things and so is a three-place relation.
We can think of predicates as one-place relations. For symbolizing relations
we use capital letters, and call them relation symbols. The same symbol can
be used to symbolize a statement (in which case it is functioning as a sentential
letter) or a relation. So we make no distinction between relation symbols
and sentential letters.
The order in which we write the individual constants after a relation
symbol makes a difference to what is expressed. If Tjb means “John is taller
than Bob”, then Tbj means “Bob is taller than John”. In specifying the
meaning of a relation symbol, we must say in what order the individual
constants are to be written. We can do this with the following device.
Rather than just saying that T stands for the relation “is taller than”, we can
say that it stands for the relation “(1) is taller than (2)”. Then in using T to
symbolize a statement, we consider what is in the place of the (1) and what
is in the place of the (2). In symbolizing “John is taller than Bob”, “John” is
in the place of the (1) and “Bob” is in the place of the (2). This signifies that
“John” comes first and “Bob” comes second. Thus we symbolize the statement
as Tjb. If instead we had taken T to stand for the relation “(2) is taller than
(1)”, then in the statement “John is taller than Bob”, “John” would occupy
the place of the (2), indicating that it comes second, and “Bob” would
occupy the place of the (1), indicating that it comes first. Then the statement
would be symbolized as Tbj. We might diagram this as follows:

John is taller than Bob.

(2) is taller than (1).

T b j

In general, if we have a relation “(1) ... (2)”, and we have some statement of
the form “b ... c”, where b and c denote individuals, then in deciding in
what order to write b and c after the relation symbol, we match up the
statement with the relation and see which individual constant corresponds
to which number. The individual constant corresponding to (1) is written
first, and the individual constant corresponding to (2) is written second.
To illustrate this with a slightly more complex example, consider the
three-place relation “is between”. For example, we might say that John is
between Bob and William. In symbolizing this statement we might let B
THE SYNTAX OF THE PREDICATE CALCULUS 104

stand for “(1) is between (2) and (3)”. Then letting j stand for “John”, b for
“Bob”, and m for “William”, we can symbolize the form of “John is between
Bob and William” as Bjbm. Had we instead let B stand for “(2) is between
(1) and (3)”, then the statement would be symbolized as Bbjm.

Exercises

A. Letting T be “(1) is taller than (2)” and B be “(1) is between (2) and (3)”,
and letting j stand for “John”, b for “Bob”, and m for “William”, symbolize
the forms of the following statements:
1. If Bob is taller than John, then it is not the case that John is taller than
Bob.
2. If John is taller than Bob, and Bob is taller than William, then John is
taller than William.
3. It is not the case that John is taller than himself.
4. If John is between Bob and William, then John is between William
and Bob.
5. If John is between Bob and William, then neither is Bob between
John and William nor is William between John and Bob.
B. Letting B be “(2) is between (1) and (3)”, symbolize the fourth and fifth
statements of the preceding exercise.

3. Quantifiers and Variables


We are still not able to symbolize the forms of statements like “All
apples are red”. For that we must introduce the concepts of a variable and
a quantifier. Consider the statement “Given any two things, either the first
is taller than the second, or the second is taller than the first”. To symbolize
the form of this statement, we need a symbolic device that will work like
the expressions “the first” and “the second” to allow us to talk about two
unspecified objects. To do this we borrow the notion of a variable from
algebra. In algebra we write things like x + y = y + x. This means “Given
any two numbers, the first added to the second is equal to the second
added to the first”.1 In other words, we use the variables x and y instead of
the expressions “the first” and “the second”. Using variables in this way,
we can paraphrase a statement like “Given any object, there is an object
taller than it” as “Given any object x, there is and object y such that y is
taller than x”. In other words, variables act much like pronouns. They
allow us to refer to the same object several times and to keep it separate

1
Notice that when we write “Given any two numbers”, we do not mean
“Given any two different numbers”. The two numbers may be equal; for
example x and y in x + y = y + x may both refer to the same number.
THE SYNTAX OF THE PREDICATE CALCULUS 105

from other objects we may be talking about. A variable will be any lower
case letter of the alphabet from t through z, with or without numerical
subscripts. For example, x, z, t2 , and w137 are variables.
We can paraphrase statements in which “every”, “all”, “any”, “some”,
and “there is” occur by using variables. “Given any two objects, if the first
is taller than the second, then it is not the case that the second is taller than
the first” can be paraphrased as “Given any two objects x and y, if x is taller
than y, then it is not the case that y is taller than x”. If we now write
variables after relation symbols just like we wrote individual constants after
them, we can partially symbolize this statement as “Given any two objects,
x and y, (Txy → ~Tyx)”. Simple statements like “Everything is red” can
also be paraphrased in this way: “Given anything x, Rx”.
Now consider how we might symbolize the form of the statement “All
apples are red”. The first step is to paraphrase it to get it into the form
“Given anything x, .. ” What can we put in the blank? “All apples are red”
means “Everything that is an apple is red”, or “Given anything if it is an
apple then it is red”. So this can be paraphrased as “Given anything x, if x
is an apple then x is red”. Letting A be “is an apple” and R be “is red”, we
can partially symbolize this as “Given anything x, (Ax → Rx)”. Now all
that remains is to introduce some way to symbolize the phrase “Given
anything x”. We will do this using an inverted “A”: “(∀x)”. Then, “Given
anything x, (Ax → Rx)” can be symbolized as (∀x)(Ax → Rx). “(∀x)” is
called the universal quantifier with respect to x. Similarly, “(∀y)” is the universal
quantifier with respect to y, and so on. Note that it makes no difference
which particular variable we use in symbolizing something. We can
symbolize “All apples are red” as either (∀x)(Ax → Rx) or (∀y)(Ay → Ry).
The reason for having more than one variable at our disposal will become
apparent when we discuss formulas that contain more than one quantifier.
In general, a statement of the form “All A are B” can be paraphrased
as “Given anything x, if x is A then x is B”. So “All A are B” can be
symbolized as (∀x)(Ax → Bx).
The universal quantifier corresponds to the expressions “all” and
“every”. It is convenient to have another kind of quantifier corresponding
to the expressions “some” and “there is”. Consider the statement “Some
apples are green”. This can be paraphrased as “There is something x, that
is a green apple”. To say that something is a green apple is just to say that
it is both green and an apple. So “Some apples are green” can be paraphrased
as “There is something x, such that (Gx & Ax)”. Next we want to introduce
a symbol for the phrase “There is something x, such that”. The symbol
“(∃x)” will be used. This is called the existential quantifier with respect to x.
“Some apples are green” can be symbolized as (∃x)(Gx & Ax). In general, a
statement of the form “Some A is B” or “There is an A that is B” can be
symbolized as (∃x)(Ax & Bx).
It should be noted that in symbolizing a statement of the form “Some
A is B” we use a conjunction, whereas in symbolizing a statement of the
form “All A are B” we use a conditional. There seems to be a strong
temptation to mix these up, and either symbolize “All A are B” as (∀x)(Ax
& Bx) or “Some A is B” as (∃x)(Ax → Bx). These are wrong for the following
reasons. First consider (∀x)(Ax & Bx). This says that everything is both an
A and a B. But that is not what we mean when we say that all A are B. We
THE SYNTAX OF THE PREDICATE CALCULUS 106

do not mean to imply that everything is an A. Thus we must write (∀x)(Ax


→ Bx) rather than (∀x)(Ax & Bx). Similarly, consider (∃x)(Ax → Bx). Recall
that (Ax → Bx) is tautologically equivalent to (~Ax ∨ Bx). What (∃x)(~Ax ∨
Bx) says is that there is something that either is not an A or is a B. This will
be true if there is something which is not an A, regardless of whether there
is anything that is a B. For example, if G is “is green” and A is “is an
apple”, then (∃x)(~Ax ∨ Gx) would be true just as long as there is something
that is not an apple. It does not require that any apples be green. Consequently,
we must symbolize “Some A is B” as (∃x)(Ax & Gx) rather than (∃x)(Ax →
Gx).
One way to understand why “All A are B” is symbolized using a
conditional but “Some A is B” is symbolized using a conjunction is to note
that “Some A is B” is equivalent to “It is not the case that all A are non-B”.
The latter would be symbolized as ~(∀x)(Ax → ~Bx). Intuitively, ~(∀x)... is
equivalent to (∃x)~... . That is, to say that it is false that everything has a
property is just to say that something lacks it. So ~(∀x)(Ax → ~Bx) is equivalent
to (∃x)~(Ax → ~Bx). By the propositional calculus, ~(Ax → ~Bx) is equivalent
to (Ax & Bx), so ~(∀x)(Ax → ~Bx) is equivalent to (∃x)(Ax & Bx). In other
words, the reason “Some A is B” is symbolized using a conjunction is that
the negation of a conditional is equivalent to a conjunction.
A formula of the form (∀x)( . . . ) is called a universal generalization
(with respect to x), and a formula of the form (∃x)( . . . ) is called an
existential generalization (with respect to x). Universal and existential
generalizations were introduced to symbolize statements involving “all”,
“every”, “some”, and “there exists”, but there are some other expressions
that can also be symbolized with their help. One such expression is “No”.
Consider the statement “No positive integer is less than itself”. This can be
paraphrased and then symbolized as follows:

No apple is blue.
It is not the case that there is a blue apple.
~(∃x)(x is an apple and x is blue)
~(∃x)(Ax & Bx)

Note that there is an alternate way of paraphrasing and symbolizing this


statement:

No apple is blue.
Every apple is non-blue.
(∀x)(if x is an apple then x is not blue)
(∀x)(Ax → ~Bx)

In general, “No A is B” can be paraphrased either as “It is not the case that
there is an A that is B” or as “Given any A, it is not a B”. The first
paraphrase produces the symbolization ~(∃x)(Ax & Bx), and the second
produces the symbolization (∀x)(Ax → ~Bx). It will be shown later that the
formulas ~(∃x)(Ax & Bx) and (∀x)(Ax → ~Bx) are equivalent, so it makes no
difference which way we symbolize a statement involving “No”.
The English word “any” behaves peculiarly. In some contexts it
THE SYNTAX OF THE PREDICATE CALCULUS 107

expresses a universal generalization, but in others it expresses an existential


generalization. For example, “If anyone was president before Jefferson, he
was president before Lincoln”, means “Everyone who was president before
Jefferson was president before Lincoln”, which is a universal generalization.
But, “If anyone was president before Jefferson, then Jefferson was not the
first president”, means, “If there was a president before Jefferson, then
Jefferson was not the first president. The latter is a conditional with an
existential antecedent. So in symbolizing English sentences containing “any”,
one must think carefully about what they mean rather than treating them
mechanically.

Exercises

Letting P stand for “was a president of the United States”, B for “(1) was
president before (2)”, and letting g stand for “Washington”, l for “Lincoln”,
and j for “Jefferson”, symbolize the forms of the following statements:
1. There was no president before Washington.
2. Washington was president before anyone who was president after
Lincoln.
3. Someone was president before Lincoln and after Jefferson.
4. Everyone who was president before Jefferson was president before
Lincoln.

4. Statements with Several Quantifiers


More than one quantifier is required for symbolizing the forms of
some statements. Consider:
Given any two things, if the first is taller than the second, then it is false
that the second is taller than the first.
We can paraphrase this as
Given any two things x and y, if Txy then ~Tyx.
This can be partially symbolized as
Given any two things x and y, (Txy → ~Tyx).
This in turn can be paraphrased as
Given anything x, given anything y, (Txy → ~Tyx)
which can then be symbolized as
(∀x)(∀y)(Txy → ~Tyx).
Consider some further examples of symbolizing the forms of statements
using formulas of the predicate calculus. Let P stand for the predicate “is a
THE SYNTAX OF THE PREDICATE CALCULUS 108

positive integer”, let E stand for “is even”, let O stand for “is odd”, let L
stand for “(1) is less than (2)”, and let I stand for “(1) is equal to (2)”. Let a1
stand for the numeral “1”, a2 for the numeral “2”, a3 for the numeral “3”,
and so on.
First consider a few examples that do not involve quantifiers. The
statement “1 is less than 2” will be symbolized as La1 a2. The statement “1 is
less than or equal to 2” can be paraphrased as “1 is less than 2, or 1 is equal
to 2”, which can then be symbolized as (La1 a2 ∨ Ia1 a2). The statement “7 is
greater than 5”, can first be paraphrased as “5 is less than 7”, and then
symbolized as La5a7 .
Now consider some statements involving quantifiers. The statement
“Every positive integer is either odd or even” can be paraphrased as “Given
anything x, if x is a positive integer than either x is odd or x is even”. This
can then be partially symbolized as (∀x)[if Px then either Ox or Ex], and
then finished as (∀x)[Px → (Ox ∨ Ex)].
The statement “There is a positive integer greater than every positive
integer” can be paraphrased as “There is something x, such that x is a
positive integer, and x is greater than every positive integer”. This in turn
can be partially symbolized as (∃x)[Px and x is greater than every positive
integer]. This can be paraphrased as (∃x)[Px and given anything y, if y is a
positive integer then x is greater than y]. This can be partially symbolized
as (∃x)[Px and (∀y)(if Py then Lyx)], and then finished as (∃x)[Px & (∀y)(Py
→ Lyx)].
The symbolization of the last statement illustrates the general pattern
that should be followed in complex cases. The process of symbolizing is
done in two steps First we introduce quantifiers and variables. Second we
symbolize the rest as in the propositional calculus, beginning with the smallest
parts and working outwards. It is illuminating to diagram this process as
follows:

There is a positive integer greater than every positive integer.

( ∃x)[x is a positive integer and x is greater than every positive integer]

(∃x)[x is positive integer and (∀y)(if y is a positive integer then x is


greater than y)]

(∃x)[ Px & (∀y)(if Py then Lyx)]

(∃x)[Px & (∀y)(Py → Lyx)]

When symbolizing a complex statement, we begin with the outermost


(leftmost) quantifier and then work inwards, introducing new quantifiers
and variables as we go. We begin with the largest parts and work down to
the smallest parts. Then in the second step, after we have introduced all of
the quantifiers in the formula, we symbolize the rest of the formula as in the
propositional calculus, beginning with the smallest parts and working
outwards to the larger parts.
The preceding example illustrates something important. Notice that
“(∀y)” comes in the middle of the formula—just where it comes in
THE SYNTAX OF THE PREDICATE CALCULUS 109

paraphrasing the original sentence. It is important to realize that the


quantifiers are not automatically placed at the front of the formula. It is a
common mistake to try to put them there.
Consider another example:
Given any positive integer, there is a greater positive integer.
In symbolizing this, the first step is to introduce quantifiers and variables,
from the outside working in:
Given any positive integer, there is a greater positive integer.
(∀x)[if Px then there is a positive integer greater than x]
(∀x)[if Px then (∃y)(Py and Lxy)]
Then we finish up by symbolizing the rest of the expression as in the
propositional calculus:
(∀x)[Px → (∃y)(Py & Lxy)]
Consider another example. The statement “Given any odd positive
integer, there is a greater even positive integer” can be symbolized as follows:
Given any odd positive integer, there is a greater even positive integer.
(∀x){if x is an odd positive integer, then there is an even positive integer
greater than x}
(∀x){if (Px & Ox) then there is an even positive integer greater than x}
(∀x){if (Px & Ox) then (∃y)[y is an even positive integer and y is greater
than x]}
(∀x){(Px & Ox) → (∃y)[(Py & Ey) & Lxy]}
If the general strategy for symbolizing statement forms in the predicate
calculus is borne in mind, the student can mechanically symbolize statements
so complex that they are hard to understand in English. Remember that the
general strategy is to begin with the outermost quantifiers and work inwards
when introducing quantifiers and variables. Then those parts that do not
involve quantifiers can be symbolized just as in the propositional calculus.
To illustrate the potency of this approach, consider an almost unintelligible
English sentence:
Everyone loves someone who doesn’t love anyone who loves himself.
Letting P symbolize “is a person” and L symbolize “(1) loves (2)”, we can
symbolize the statement as follows:
(∀x) if x is a person then x loves someone who doesn’t love anyone
who loves himself.
(∀x) if x is a person then there is someone y such that Lxy and y
doesn’t love anyone who loves himself.
(∀x){Px → (∃y){(Py & Lxy) & y doesn’t love anyone who loves himself}
(∀x){Px → (∃y)[(Py & Lxy) & there is no z such that z is a person and y
THE SYNTAX OF THE PREDICATE CALCULUS 110

loves z and z loves himself]}


(∀x){Px → (∃y)[(Py & Lxy) & ~(∃z)(z is a person and y loves z and z
loves himself)]}
(∀x){Px → (∃y)[(Py & Lxy) & ~(∃z)((Pz & Lyz) & Lzz)]}

Thus it is straightforward to symbolize the sentence even though we have


trouble grasping its meaning. Symbolizing it may actually help us in
understanding it.

Exercises

A. Letting P stand for “was a president of the United States”, B for “(1) was
president before (2)”, D for “(1) died before (2)”, A for “(1) was born before
(2)”, and P2 for “is a person”, and letting g stand for “Washington”, l for
“Lincoln”, and j for “Jefferson”, symbolize the forms of the following
statements:
1. Given any two presidents, if the first was born before the second,
then the first died before the second.
2. Some presidents were born before others, and some died after others.
3. There is at least one president who was born before another president
and died after him.
4. There were no two presidents between Jefferson and Lincoln such
that one was born before the other but died after him.
5. There is a president who was president both before and after another
president.
6. Given any three past presidents, if the first died before the second,
and the second died before the third, then the first died before the
third, but it is not the case that given any three presidents, if the first
was born before the second, and the second was president before the
third, then the third died after the second.
7. Given any two presidents, if the first died before Jefferson, and Lincoln
died before the second, then if the second was born before the first,
then the second was president before the first and died after
Washington but before Jefferson.
8. No president was president both before and after two presidents
who in turn were each president both before and after each other.
B. Letting D stand for “is a dog”, A for “is a day”, S for “is a stone”, L for
“is an angel”, R for “is a time when it rains”, B for “is rolling”, P for “is a
time when it pours”, F for “is a fool”, M for “is moss”, H for “(1) has (2)”, U
for “(1) rushes into (2)”, G for “(1) gathers (2)”, and E for “(1) fears to tread
in (2)”, the following formulas symbolize familiar sayings. Express these
sayings in English.
THE SYNTAX OF THE PREDICATE CALCULUS 111

1. (∀x)(Rx → Px)
2. (∀x)[Dx → (∃y)(Ay & Hxy)]
3. (∀x)[(Sx & Bx) → ~ (∃y)(My & Gxy)]
4. (∀x)(∀y){[(∃z)(Lz & Ezy) & Uxy] → Fx}

5. Symbolizing Relations
Thus far we have only been concerned with symbolizing the forms of
statements, but the same thing can be done with relations. Some relations
can be viewed as constructed out of other simpler relations. For example,
the relation “Either (1) loves (2), or (2) loves (1)” is constructed out of the
simpler relation “(1) loves (2)”. Taking the relation symbol L to stand for
the simpler relation “(1) loves (2)”, we can symbolize the form of the complex
relation by writing [L(1)(2) ∨ L(2)(1)]. We can then go a step further and
replace the numbers in parentheses by variables. The result is the formula
(Lxy ∨ Lyx), which gives us the logical form of the complex relation.
Consider a more complicated example:
If (1) strikes (2) in the presence of (3), then (3) will think poorly of (1),
but (2) will be embarrassed and will think poorly of himself.
Letting S stand for “(1) strikes (2) in the presence of (3)”, letting T stand for
“(1) thinks poorly of (2)”, and letting E stand for “is embarrassed”, we can
symbolize the complex relation as
If S(1)(2)(3), then T(3)(1), but E(2) and T(2)(2)
{S(1)(2)(3) → [T(3)(1) & (E(2) & T(2)(2))]}
When we replace the numbers by variables we get
{Sxyz → [Tzx & (Ey & Tyy)]}
In replacing the numbers by variables it makes no difference which variables
we choose. We could have symbolized the above relation equally well as
{Szyx → [Txz & (Ey & Tyy)]}.
Sometimes we must use quantifiers in symbolizing the form of a relation.
The relation
(1) is the father of the mother of (2)
can be paraphrased as
There is something, x, such that (1) is the father of x and x is the mother
of (2).
Letting F stand for “(1) is the father of (2)” and M stand for “(1) is the
mother of (2)”, we can symbolize this as
(∃x)(F(1)x & Mx(2)).
THE SYNTAX OF THE PREDICATE CALCULUS 112

Replacing “(1)” by y and “(2)” by z we obtain


(∃x)(Fyx & Mxz).
Notice that the variables we substitute for “(1)” and “(2)” must be
different from the variable we use with the quantifier. Had we used x
again for “(1)” and y for “(2)”, the result would have been the formula
(∃x)(Fxx & Mxy). This means “There is something that is its own father,
and it is the mother of (1)”. Each time we introduce a variable to serve a
new purpose in symbolizing the form of a relation, it must be distinct from
any we have used before.

Exercises

A. Symbolize the forms of the following relations, letting M stand for “(1)
is married to (2)”, W stand for “(1) is a wife of (2)”, F stand for “(1) is the
father of (2)”, P stand for “(1) is a parent of (2)”, B stand for “(1) is a brother
of (2)”, S stand for “(1) is a sister of (2)”, and D stand for “(1) is a daughter
of (2)”.
1. (1) is a grandfather of (2).
2. (1) is the paternal grandfather of (2).
3. (1) is a sister-in-law of (2). (A sister-in-law is a wife of a brother or a
sister of a spouse.)
4. (1) is the paternal grandfather of (2)’s maternal grandmother.
5. (1) is a cousin of (2). (A cousin is an offspring of a sibling of a parent.)
6. (1) is a niece of (2). (A niece is a daughter of one's brother or sister or
daughter of the brother or sister of one's spouse.)
7. (1) is a nephew of (2). (A nephew is a son of one's brother or sister or son
of the brother or sister of one's spouse.)
8. (1) is an uncle of (2). (An uncle is a brother of a parent or the spouse of a
sister of a parent.)
9. (1) is a cousin of a nephew of an uncle of (2).
10. A cousin of a sibling of (1) is a nephew of (2) and a grandmother of
(3).

6. Of Freedom and Bondage


When we symbolize the form of a statement and introduce a quantifier
together with one or more occurrences of the corresponding variable, we
say that the quantifier binds those occurrences of the variable. Therefore,
the quantifier “(∀x)” binds both occurrences of x in (∀x)(Ax → Bx). Similarly,
THE SYNTAX OF THE PREDICATE CALCULUS 113

the quantifier “(∀x)” binds all occurrences of x and the quantifier “(∃y)”
binds all occurrences of y in the formula (∀x)(Ax → (∃y)Rxy). It should be
noticed that different occurrences of the same variable can be bound by
different quantifiers. The statement “Either all apples are red or all apples
are green” can be symbolized as
[(∀x)(Ax → Rx) ∨ (∀x)(Ax → Gx)].
Here the first and second occurrences of x are bound by the first quantifier,
and the third and fourth occurrences are bound by the second quantifier.
(We don’t count the variable in the quantifier itself as an occurrence.) To
take a more complex case, in
(∀x)[(Ax → Bx) & (∃x)Ax]
the universal quantifier binds the first two occurrences of x, and the existential
quantifier binds the third occurrence of x. An occurrence of a variable is
always bound by the innermost quantifier that can bind it. Thus in the above
example the third occurrence of x is bound by the existential quantifier
rather than the universal quantifier. If the existential quantifier were omitted
then all three occurrences of x would be bound by the universal quantifier.
In general, in a formula of the form (∀x)( . . . x . . . ) or (∃x)( . . . x . . . ), the
quantifier binds all occurrences of x within the formula that are not already
bound by another quantifier.
The formula that immediately follows a quantifier and in which any
occurrences of the variable of that quantifier not previously bound are bound
by the quantifier, is called the scope of the quantifier. So, in the formula
[(∀x)(Ax → Rx) ∨ (∀x)(Ax → Gx)], (Ax → Rx) is the scope of the first
quantifier, and (Ax → Gx) is the scope of the second quantifier. In the
formula (∀x)(Ax → (∃y)Rxy), (Ax → (∃y)Rxy) is the scope of “(∀x)”, and
Rxy is the scope of “(∃y)”. An occurrence of a variable is bound by a quantifier
if, and only if, the quantifier contains that variable, and the occurrence in question
is within the scope of that quantifier and not already bound by another quantifier.
Notice that in a formula we get by symbolizing the form of a statement,
each occurrence of a variable must be bound by some quantifier. This is
because we only introduce variables along with quantifiers that bind them.
Variables only result from symbolizing English quantifiers like “all”, “each”,
“every”, “some”, and so on. On the other hand, when we symbolize relations,
the variables that we put in place of the numbers in parentheses are not
bound by any quantifiers. If an occurrence of a variable in a formula is not
bound by any quantifier, it is called a free occurrence. For example, the
occurrence of z in (x)(∃y)(Fxy ↔ Hxyz) is free. Hence, the difference between
the formulas that result from symbolizing the forms of statements and those
that result from symbolizing the forms of relations is that the former never
have free occurrences of variables, and the latter always have fee occurrences
of variables.
A formula of the predicate calculus that does not have any free
occurrences of variables is called a closed formula. So we always get closed
formulas when we symbolize the forms of statements. A formula that
contains free occurrences of variables is called an open formula. Open formulas
symbolize relations.
THE SYNTAX OF THE PREDICATE CALCULUS 114

Exercises

A. In the following formulas, indicate the scope of each quantifier, and


which occurrences of variables are bound by each quantifier:
1. (∀x)(Ax → Bx)
2. (∀x)(Axb → (∃y)Hxy)
3. [Rab ∨ (∀x)(∃y)Hxy]
4. (∀x)[Hxa → {(Fax ↔ Hbx) → ~(∃x)Hxx}]
5. (∀x)(∀y)(∀z)Hxyz
6. [(∀x)(Ax → Bx) ∨ ~(∃x)(Ax & ~Bx)]
7. (∀x){Hx → (∀x)[Hx → (∀x)(Hx → (∀x)Hx)]}

B. For each of the following formulas, tell which (if any) occurrences of
variables are free:
1. Bxa
2. (∀x)(Fxa → Bxy)
3. [(∃x)(Gx & Hyz) ∨ Hxx]
4. (∀x)[Hxa ↔ (∃y)(∃z)(Hyx → ~Hzz)]
5. (∀x)(Hx1 → Bxx)
6. [(∀x)(∃y)(Bxyz ∨ ~ Bzyx) ∨ ~(Bxyz ∨ ~Bzyx)]

7. Universes of Discourse
When we write (∀x)Fx, we are saying that everything in the universe
has the property F. This is sometimes expressed by saying that the variables
in quantifiers “range over” everything in the universe. However, we are
often interested in only a subset of all the objects in the universe. For
example, when we write “x + y = y + x” in mathematics, we are implicitly
supposing that our variables range only over numbers. We could express
this precisely by taking N to symbolize “is a number” and writing:
(∀x)(∀y)[(Nx & Ny) → x + y = y + x].
However, when we are only interested in talking about numbers, it is more
convenient to eliminate the use of N and just write:
(∀x)(∀y) x + y = y + x
understanding that the variables just range over numbers. When we do
this we say that we are employing a restricted universe of discourse. If our
THE SYNTAX OF THE PREDICATE CALCULUS 115

universe of discourse consists exclusively of numbers, then “(∀x)” means


“Given any number x” and “(∃x)” means “There is a number x such that”.
This can produce formulas that are significantly shorter and easier to read.
For example, without using a restricted universe of discourse, we must
symbolize “For every number, there is a larger number” as
(∀x)[Nx → (∃y)(Ny & Lyx)].
But if we restrict our universe of discourse to numbers, this can be written
much more simply as (∀x)(∃y)Lyx.
Restricted universes of discourse are particularly useful in mathematics,
but they can be useful elsewhere as well. For instance, in some contexts we
may only be talking about persons. If we symbolize “Everyone is loved by
someone” without using a restricted universe of discourse, we will get
(∀x)[Px → (∃y)(Py & Lyx)],
but if we restrict our universe of discourse to persons we can again write
(∀x)(∃y)Lyx.
In the predicate calculus, we are allowed to take any class C of things
as our universe of discourse. C could be the class of numbers, or the class
of persons, or anything else that is convenient. When we do this, “(∀x)”
means “Given anything x in the class C” and “(∃x)” means “There is something
x in the class C such that”.

Exercises

Taking the class of human beings as your universe of discourse, and letting
M stand for “(1) is married to (2)”, W stand for “(1) is a wife of (2)”, F stand
for “(1) is the father of (2)”, P stand for “(1) is a parent of (2)”, B stand for
“(1) is a brother of (2)”, S stand for “(1) is a sister of (2)”, a stand for
“Arthur”, and b stand for “Bartholomew”, symbolize the forms of the
following statements. You can assume that Arthur and Bartholemew are
are humans and that parents and siblings of humans are humans.
1. It is false that anyone is both the father of Arthur and the mother of
Bartholomew.
2. Arthur has a mother who is married to his father.
3. Arthur is the paternal grandfather of Bartholomew.
4. Arthur is a sibling of someone.
5. Everyone who is married to anyone is the offspring of someone.
6. One of Arthur’s grandfathers is a brother of one of Bartholomew’s
grandmothers.
7. Anyone who is married to someone’s brother is someone’s sister-in-
law.
8. No one is the father of a brother of his mother.
THE SYNTAX OF THE PREDICATE CALCULUS 116

8. Formulas of the Predicate Calculus


We have been using formulas of the predicate calculus to symbolize
statements and relations. However, we have yet to give a precise definition
of the concept of a formula of the predicate calculus. We can give such a
definition by making it analogous to the definition of a formula of the
propositional calculus. We have already specified which symbols are
variables, individual constants, and relation symbols. Remember that
relations symbols also function as sentential letters. Then just as in the
propositional calculus, we define atomic formulas to be formulas that contain
no logical symbols, and we define molecular formulas to be formulas that
do contain logical symbols. We now have a larger class of logical symbols,
because in the predicate calculus we include ∀ and ∃ as logical symbols.
As in the propositional calculus, sentential letters are atomic formulas.
They are the simplest atomic formulas. However, in the predicate calculus
there are other atomic formulas as well. These include expressions like Fxy
or Gcd. In general, an atomic formula of the predicate calculus consists of a
relation symbol followed by any combinations of zero or more variables or
individual constants. So, for example, the following are all atomic formulas:
P, Q, A13, F1 x, F3xxx, G3xyz, A237 ab, B4xaxa
Then all other formulas of the propositional calculus can be constructed by
successive applications of the following rules:
1. An atomic formula is a formula.
2. If A is any formula, then ~A is a formula.
3. If A and B are any formulas, then (A & B) is a formula.
4. If A and B are any formulas, then (A ∨ B) is a formula.
5. If A and B are any formulas, then (A → B) is a formula.
6. If A and B are any formulas, then (A ↔ B) is a formula.
7. If A is any formula and x is any variable, then (∀x)A is a formula.
8. If A is any formula and x is any variable, then (∃x)A is a formula.

The notions of free and bound occurrences of variables can also be


made more precise. In a formula of the form (∀x)A or (∃x)A constructed
by Rules 7 or 8, A is the scope of the quantifier. It should be observed that
by requiring A to be a formula we ensure that in any given formula, there
can only be one scope for each quantifier. For example, in ((∃x)Fx ∨ P) the
scope of the quantifier must be Fx. It cannot be Fx ∨ P because this is not a
formula (it lacks parentheses). An occurrence of a variable in a formula is
bound if, and only if, it is within the scope of a quantifier containing that
variable. An occurrence of a variable is free if, and only if, it is not bound.
A formula that contains no free occurrences of variables is a closed formula.
An occurrence of a variable is bound by a particular quantifier if, and only
if, the variable in the quantifier is the same as the variable in question, and
the quantifier is the innermost quantifier whose scope contains the occurrence
THE SYNTAX OF THE PREDICATE CALCULUS 117

in question.
A peculiarity of this definition is that by Rules 7 or 8 a quantifier can
be appended to a formula even if that formula contains no free occurrences
of the variable of the quantifier. Such a quantifier is said to be vacuous. We
could have framed Rules 7 and 8 in such a way as to preclude vacuous
quantifiers, but that would have made them more complicated. It is simpler
to allow vacuous quantifiers, as long as it is clearly understood what they
mean. A formula with a vacuous quantifier will be taken to mean the same
thing as the corresponding formula without the quantifier. Thus, for example,
the formula (∀x)((∃y)Fy → (∃z)Gz) means the same thing as ((∃y)Fy →
(∃z)Gz). In other words, a vacuous quantifier leaves the meaning of a
formula unchanged.

Exercises

Which of the following are formulas of the predicate calculus? Which are
closed formulas and which are open formulas?
1. (∀x)(∀y)(Fxy → Fyx)
2. (P → (∃x)Gxy)
3. ((∀x)(∃y)(∀z)Gxy & Fyz)
4. (∃z)(∀x)Fxz → (∀x)(∃y)Gxy
5. (∃x) ~(P & (∀x)(Fx → Gx))
6. (∀x)(∀y)((Bxa → Bxy) ∨ ~(Hxyz & Bzy))
7. (∀x)(Fx) → (∃y)(Gxy → Fy)
8. ((∀w)Pw ↔ ~(∃z)(Wp & Pz))
9. (Wa & ~Wa)
10. ((P & Q) → R)

Practice Problems

Symbolize the forms of the following statements, using the symbols indicated:
1. If someone is dead, there is a murderer in the house. [D: “is dead”;
P: “is a person”; M: “is a murderer”; H: “is in the house”]
2. If someone is dead, Jones killed him. [D: “is dead”; P: “is a person”;
K: “(1) killed (2)”; j: Jones]
3. Any student in some course John is in will flunk it if John does. [S:
“is a student”; C: is a course”; I: “(1) is in (2)”; F: “(1) flunks (2)”, j:
THE SYNTAX OF THE PREDICATE CALCULUS 118

John]
4. If all members of the ping pong team are incapacitated, then if Castro
invites them to visit Cuba none of them will be able to go. [P: “is a
member of the ping pong team”; I: “is incapacitated”; V: “(1) invites
(2) to visit (3)”; A: “(1) is able to go to (2)”; a: Castro; c: Cuba]
5. If the President gets his way, no congressman will dare vote against
the budget unless he (the congressman) is either retiring or from
California. [G: “gets his way”; C: “is a congressman”; D: “dares to
vote against the budget”; R: “is retiring”; F: “is from California”; p:
the President]
6. No one who has been to Xanadu and seen the stately pleasure dome
now remembers how to get there. [P: “is a person”; B: “(1) has been
to (2)”; R: “(1) remembers how to get to (2)”; a: Xanadu; d: the stately
pleasure dome]
7. If there are no prime numbers between 13 and 17, then if any number
between 13 and 17 is divisible by 13, it is also divisible by 4. [N: “is a
number”; P: “is prime”; L: (1) is less than (2)”; D: “(1) is divisible by
(2)”; a3 , a4, a13, a17]
8. If anyone steals the royal crown and melts it down, no one will miss
it unless the king does. [S: “(1) steals (2); P: “is a person”; M: “(1)
melts (2) down”; I: “(1) misses (2)”; k: the king; c: the royal crown]
9. No one respects anyone who does not respect anyone. [P: “is a
person”; R: “(1) respects (2)”]
10. Anyone who is taller than someone is taller than someone who isn’t
taller than anyone. {P: “is a person”; T: “(1) is taller than (2)”]
6
The Semantics of the
Predicate Calculus

1. The Rudiments of Set Theory


In the propositional calculus interpretations were defined to be
assignments of truth values to sentential letters. We then defined a valid
formula (a tautology) to be a formula true relative to every interpretation,
and showed that a formula of the propositional calculus is a tautology if,
and only if, it expresses a formally necessary statement form. We want to
do something similar for the predicate calculus. However, because of the
greater expressive power of the predicate calculus, interpretations must do
more than assign truth values to sentential letters. They must also interpret
relation symbols, individual constants, and quantifiers. This can be done
set-theoretically. For example, quantifiers will be interpreted by giving a
set of objects which constitutes the universe of discourse. To make this
precise, the student must have some acquaintance with set theory. This
section will discuss the rudiments of set theory, and then in section two we
will return to the task of defining the concept of an interpretation for the
predicate calculus.
First, what is a set? This is a question that we can only answer by
giving synonyms. The terms “set”, “class”, “collection”, “group”, and
“ensemble” can all be used interchangeably. A set may be defined to be
any collection of objects. For example, we can talk about the set of all
integers between one and seventeen. This set consists of the integers from
two through sixteen, inclusive. Or we may talk about the set of all desks in
a room, or the set of all buildings on the campus, or the set of all past
Presidents of the United States, and so forth.
A set is a set of objects. The objects in a set are called the members or
elements of the set. Given any set, A and any object x, we abbreviate “x is a
member of A” by writing x∈A. Thus, for example, if A is the set of all past
Presidents of the United States, and k stands for “John F. Kennedy”, we can
write k∈A. Or, if N is the set of integers between one and sixteen, then we
can write 6∈N.
Perhaps the most important property of sets is that they are extensional.
This means that the identity of a set is determined entirely by what objects
it has as members—if two sets have exactly the same members, then they
are the same set. For example, the set of all integers between one and four
is identical to the set of the two integers two and three, because these sets
have the same members. Or the set of featherless bipeds is the same as the
set of things that are either human beings or plucked birds. Using the
symbols of the predicate calculus, we can express what it means to say that
sets are extensional as follows: Given any two sets A and B,
THE SEMANTICS OF THE PREDICATE C ALCULUS 120

[(∀x)(x∈A ↔ x∈B) → A = B]
Because the identity of a set is determined by its members, one convenient
way to refer to a set having a small number of members is by listing its
members. The customary way to do this is to enclose the list of members in
braces, { and }. We can, for example, refer to the set of all integers between
one and four by writing {2, 3}. It makes no difference in what order we list
the members of the set. If the members are the same, the set is the same. So
{2, 3} and {3,2} are the same set, and {1,2, 3}, {1, 3,2}, {2,1, 3}, {2, 3,1}, {3,1,2},
and {3,2,1} are the same set.
Sometimes a set will have just one member. For example, we can talk
about the set of all integers between one and three, which is the set whose
only member is two. This set can be referred to as {2}. A set having only
one member is called the unit set of that object. Thus {2} is the unit set of
two.
Sometimes a set will have too many members to make it practical to
list all of them. A set might even have an infinite number of members, like
the set of all integers. When this happens we cannot refer to the set by
simply listing the members, but we can adapt the above notation to take
care of this. If for some predicate A we want to talk about the set of all
objects that are A (such as the set of all past presidents), we can simply
write {x  x is A}. This is read “the set of all objects, x, such that x is A”.
Using these symbols the set of all past presidents is {x  x is a past president}.
There is one set that deserves special mention because of its somewhat
paradoxical-seeming nature. This is the set that has no members. This is
called the empty set, and is designated by the symbol ∅. The empty set can
be introduced in many different ways. For example, ∅ = {x  x both is and
is not an integer} because there is nothing that both is and is not an integer.
Notice that there can only be one empty set—this follows from the
extensionality of sets. If we had two sets neither of which had any members,
then they would have the same members (that is, (∀x)(x∈A ↔ x∈B) would
be true), and so they would be the same set.
A set that is just the opposite of the empty set is the universal set. The
universal set is the set containing everything. It contains all numbers, material
objects, people, and so on. The universal set is denoted by U. We can
define it as U = {x  x either is or is not an integer} because everything either
is or is not an integer.
The order in which we list the members of a set makes no difference
to the identity of the set. But sometimes we would like it to make a difference.
This leads to the concept of an ordered-set, or a sequence. A sequence is a list
of objects having a definite order. For example, we may talk about the
sequence of integers from one to ten, listed in order from smallest to largest.
This will be distinct from the sequence of integers from one to ten listed in
order from largest to smallest.
Just as { and } are used to talk about sets, 〈 and 〉 are used to talk about
sequences. Thus, the sequence of integers from one to ten listed in order
from smallest to largest will be the sequence 〈1,2, 3, 4, 5, 6, 7, 8, 9,10〉, and
THE SEMANTICS OF THE PREDICATE C ALCULUS 121

the sequence of integers from one to ten listed in the reverse order will be
〈10, 9, 8, 7, 6, 5, 4, 3,2,1〉. These are different sequences. Unlike sets, sequences
are not extensional. Whereas the identity of a set is determined solely by its
members, the identity of a sequence is determined by the combination of its
members and their order.
We will frequently want to talk about sequences of fixed length. A
sequence of two objects will be called an ordered pair. Thus 〈1,2〉, 〈John
Kennedy, F. Scott Fitzgerald〉, 〈Hiroshima, √ 2〉 are all ordered pairs. Similarly,
a sequence of three objects is called an ordered triple, and a sequence of four
objects is called an ordered quadruple. Above a certain point we run out of
names like this, so we use another kind of name. An ordered pair can also
be called an ordered two-tuple, and ordered triple can be called an ordered
three-tuple, and so on. This terminology will work for a sequence of any
length. A sequence of 137 objects can be called an ordered 137-tuple. In
general, given any number n, we can talk about sequences of n objects and
call them ordered n-tuples.
There is a close relationship between sets and relations. The simplest
relations are one-place relations, or predicates. Corresponding to each
predicate is the set of all objects to which that predicate can be truly ascribed.
For example, corresponding to the predicate “is a bachelor” there is the set
of all bachelors; that is {x  x is a bachelor}. And corresponding to the
predicate “is an integer” there is the set of all integers; that is, {x  x is an
integer}. In other words, given any predicate A, there corresponds to it {x 
x is A}. This set is called the extension of the predicate A. An object is a
member of the extension of a predicate if, and only if, that predicate can
be truly ascribed to that object.
Next consider two-place relations, such as “(1) is the brother of (2)”.
There is also a set corresponding to this, but now it is a set of ordered pairs.
For example, corresponding to the relation “(1) is the brother of (2)” is the
set of all ordered pairs for which the first element of the pair is the brother
of the second element of the pair. Or corresponding to the relation “(1) is
an integer less than (2), and (2) is an integer less than four”, we have the set
of all ordered pairs of integers such that the first integer in each pair is less
than the second integer, and the second integer is less than four. This is just
the set {〈1,2〉, 〈1, 3〉, 〈2, 3〉}. In general, given a two-place relation “(1) ... (2)”
(where we fill something in for the blank), we can talk about the set of all
ordered pairs which are such that the first member of each pair stands in
this relation to the second member of that pair. This set can be referred to
as “the set of all ordered pairs, 〈x, y〉, such that x ... y”, and can be symbolized
as {〈x, y〉  x ... y}. Corresponding to the relation “(1) is the brother of (2)” is
{〈x, y〉  x is the brother of y}. This set of ordered pairs is called the extension
of the relation. The extension of the relation “(1) is an integer between (2)
and (3), and (2) is greater than four and (3) is less than nine” is the set {〈6, 5,
7〉, 〈6, 5, 8〉, 〈7, 5, 8〉, 〈7, 6, 8〉}. Given an n-place relation, the elements of a
particular ordered n-tuple of objects stand in this relation to one another
if, and only if, the ordered n-tuple is a member of the extension of the
relation.
THE SEMANTICS OF THE PREDICATE C ALCULUS 122

In the next section the concept of the extension of a predicate or relation


will be used to analyze the conditions under which a statement whose form
can be symbolized in the predicate calculus is true.

Exercises

A. Which of the following are the same (equal)?


1. {1,2, 3} 4. 〈3,1,2〉 7. {3,2, 3}
2. {3,2} 5. {2,(5 – 2),1} 8. {1,2,2,1, 3}
3. {3,1,2} 6. 〈1,2, 3〉 9. {x  x is an integer
between 0 and 4}

B. What are the extensions of the following predicates and relations?


1. “is a positive integer smaller than 7”
2. “is (or was) president of the United States after Carter”
3. “(1) was president after Carter, but before (2)”
4. “(1), (2), and (3) are positive integers smaller than four, and (1) is the
sum of (2) and (3)”

2. Interpretations
Interpretations must contain enough information to enable us to
determine the truth values of formulas relative to the interpretations. For
this purpose, an interpretation must specify a universe of discourse for the
quantifiers, and it must interpret the sentential letters, relation symbols,
and individual constants. The universe of discourse will be called the domain
of the interpretation. In the propositional calculus, it proved unnecessary to
specify the entire meaning of a sentential letter. For the purpose of computing
truth values, it was sufficient to simply assign truth values to the sentential
letters. That remains true in the predicate calculus, so we will take an
interpretation of the predicate calculus to assign truth values to sentential
letters. Similarly, an interpretation will assign a denotation to each individual
constant. The denotation of an individual constant is the object it denotes
or names.
A one-place relation symbol is used to express a predicate. The statement
symbolized by Fc is true if, and only if, the predicate expressed by F can be
correctly ascribed to the object denoted by c. But as we saw in section one,
that condition is satisfied if, and only if, the object denoted by c is in the
extension of the predicate expressed by F. Thus for the purpose of determining
the truth value of Fc, it suffices to know the extension of F and the denotation
of c. We do not have to know the “meaning” of F, in the sense of knowing
THE SEMANTICS OF THE PREDICATE C ALCULUS 123

what predicate it expresses.


This observation can be extended to n-ary relation symbols. For
example, consider the atomic formula Hbc. Suppose b denotes an object b
and c denotes an object c. Then Hbc is true if, and only if, 〈b,c〉 is in the
extension of the relation expressed by H. So for the purpose of determining
the truth value of Hbc, it suffices to know the extension of H and the
denotations of b and c.
It will turn out that something similar is true for arbitrary formulas of
the propositional calculus. All we have to know to evaluate their truth
values are the universe of discourse, the denotations of individual constants,
the extensions of relation symbols, and the truth values of sentential letters.
Accordingly, we will take an interpretation of the predicate calculus to
specify just these four items:
An interpretation of the predicate calculus is an assignment of the following:
1. A set of individuals to constitute the universe of discourse (this
is the domain of the interpretation); and
2. denotations in the domain for all individual constants; and
3. extensions in the domain for all relation symbols; and
4. truth values for all sentential letters.
Note that the denotation of an individual constant must be chosen from the
domain of the interpretation. Similarly, the extension of a relation symbol
must consist of objects from the domain. For example, the extension assigned
to a two-place relation must be a set of ordered pairs of objects in the
domain. This requirement should be emphasized, because it is a common
mistake to overlook it. If it were not required that the denotations and
extensions be chosen from the domain, formulas would not have their
intended meaning. For example, the formula [(∀x)Fx → Fc] means “If
everything in the domain, including c, is F, then c is F”. So we must ensure
that the denotation of c is in the domain. Note also that because all individual
constants must be assigned denotations, it follows that there must be at
least one thing in the domain. That is, the domain cannot be empty.
Notice that whenever we are given a universe of discourse and
“meanings” for the relation symbols and individual constants in a closed
formula we can construct a corresponding interpretation. The domain will
be the universe of discourse, and the extensions and denotations assigned
to the relation symbols and individual constants are those determined by
their meaning. In constructing an interpretation, we simply abstract from
the meaning those features that are necessary in order to determine truth
values.

3. Truth Rules
To compute the truth value of a formula relative to an interpretation,
we need truth rules, so let us turn to the task of formulating such rules. The
first thing to observe is that we cannot talk about the truth or falsity of an
open formula—only of closed formulas. Open formulas symbolize relations
THE SEMANTICS OF THE PREDICATE C ALCULUS 124

rather than statements, and only statements are true or false. For instance,
we cannot ask whether the relation (1) is the brother of (2) is true. That
question makes no sense. Truth rules can only be applicable to closed
formulas.
The truth rules employed in the propositional calculus are still correct
for the predicate calculus. However, because of the greater expressive power
the predicate calculus, those rules are not sufficient for computing truth
values for all formulas. The simplest cases they omit are those of atomic
formulas involving relation symbols. To compute truth values for these
formulas we need the following truth rules:
(1) If F is a relation symbol and c is an individual constant, Fc is true
relative to an interpretation if and only if the interpretation assigns an
object c as the denotation of c, a set F of objects as the extension of F,
and c∈F.
(2) If F is a relation symbol and c 1,...,c n are individual constants, Fc1 ...cn
is true relative to an interpreation if and only if the intepretation
assigns objects c1,...,cn as the denotations of c1 ,...,cn, a set F of ordered
n-tuples as the extension of F , and 〈c1,...,cn 〉∈F.
For example, suppose we use H to express the mathematical relation x+y =
z, and restrict our domain to the positive integers less than 4. So the
domain is the set {1,2,3}, and the extension of H is {〈1,1,2〉,〈1,2,3〉,〈2,1,3〉}.
Then if the denotation of b is 1, c is 2, and d is 3 (so Hbc expresses the
statement “1+2 = 3”), Hbc is true if, and only if, 〈1,2,3〉 is in the extension of
H. This condition is satisfied, so Hbc is true relative to this interpretation.
We will retain the following rules from the propositional calculus:
(3) An atomic closed formula that is a sentential letter is true under an
interpretation if, and only if, the interpretation assigns the truth value
“true” to it.
(4) If A is any closed formula, ~A is true if, and only if, A is false.
(5) If A and B are any closed formulas, (A & B) is true if and only if both
A and B are true.
(6) If A and B are any closed formulas, (A ∨ B) is true if and only if
either A or B (or both) are true.
(7) If A and B are any closed formulas, (A → B) is true if and only if
either A is false or B is true.
(8) If A and B are any closed formulas, (A ↔ B) is true if and only A and
B have the same truth value, i.e., either both are true or both are false.
Using the preceding truth rules we can determine the truth value of
any closed formula that does not contain quantifiers. For example, suppose
we want to know the truth value of the closed formula [P ↔ (Qa & ~Fab)]
under the interpretation that assigns the domain {1,2, 3}, the denotation 1 to
THE SEMANTICS OF THE PREDICATE C ALCULUS 125

a and the denotation 2 to b, the extension {2, 3} to Q, and the extension


{〈1,2〉, 〈2, 3〉} to F, and the truth value “true” to P. Then by Rule 3, P is true
under the interpretation. The denotation of a is 1, and this is not a member
of the extension of Q, so by Rule 1, Qa is false under the interpretation. The
ordered pair consisting of the denotations of a and b is 〈1,2〉. This ordered
pair is a member of the extension of F, so by Rule 2 Fab is true under the
interpretation. Then by Rule 4, ~Fab is false under the interpretation. Both
Qa and ~Fab are false, so by Rule 5, (Qa & ~Fab) is false under the
interpretation. P is true, so then by Rule 8 the whole closed formula [P ↔
(Qa & ~Fab)] is false under the interpretation.
We have yet to construct rules for determining whether universal
generalizations and existential generalizations are true. These rules are
more complicated than the above. To see how these rules must read, let us
begin by looking at a simple case—the formula (∀x)Fx. This formula is true
if, and only if, F is true of everything in the domain. That is, (∀x)Fx says
that everything in the domain is an F. One way of making this more
precise is the following. Let c be an arbitrarily chosen individual constant.
Then (∀x)Fx is true if, and only if, Fc would be true regardless of what (in
the domain) we let the denotation of c be. If the domain is {1,2, 3}, then
(∀x)Fx is true if, and only if, each of 1,2, and 3 are in the extension of F.
And the latter will be true provided that Fc would be true regardless of
which of 1,2, or 3 we assign as the denotation of c.
To formulate a precise truth rule for universal generalizations, we
need some way of talking about the result of varying the denotations of
individual constants. Given one interpretation, let us say that another
interpretation is an c-variant of the first interpretation if, and only if, the
second interpretation is exactly like the first interpretation except possibly
for the denotation it assigns to c. Then to talk about different c-variants of
an interpretation is just a way of letting the denotation of c vary over the
different elements of the domain. If we consider an interpretation that
assigns the domain {1,2, 3} and assigns the extension {1, 3} to F, there will be
three possible c-variants of this interpretation. These will be those
interpretations that assign the same domain to the quantifiers and the same
extension to F, and that assign the denotations 1,2, or 3 to c. Then to say
that Fc would be true regardless of what we take the denotation of a to be
means that Fc is true under each of the c-variants of the original interpretation.
We can also talk about b-variants of an interpretation, d-variants, and
so on for each individual constant. These will be those interpretations that
are exactly like the initial interpretation except that they also assign a
denotation to the individual constant in question.
W can formulate a preliminary version of a truth rule for universal
generalizations as follows:
A universal generalization (∀x)Fx is true under an interpretation if, and
only if, Fc is true under every c-variant of the initial interpretation.
To apply the analogous rule to a more complex universal generalization,
e.g., (∀x)[Fx ↔ (Gx ∨ Ha)], we choose some individual constant that does
THE SEMANTICS OF THE PREDICATE C ALCULUS 126

not occur in the formula, e.g., c, and construct the formula [Fc ↔ (Gc ∨
Ha)]. Then (∀x)[Fx ↔ (Gx ∨ Ha)] is true under an interpretation if, and
only if, [Fc ↔ (Gc ∨ Ha)] is true under every c-variant of the interpretation.
Notice that it makes no difference which individual constant we choose,
just as long as it does not already occur in the formula.
The formula [Fc ↔ (Gc ∨ Ha)] is obtained by substituting c for every
occurrence of x in [Fx ↔ (Gx ∨ Ha)]. A complication arises from the fact
that a formula can contain more than one quantifier binding the same variable.
For example, in computing the truth value of (∀x)[Fx ↔ (Gx ∨ (∃x)Hx)], we
would construct the formula [Fc ↔ (Gc ∨ (∃x)Hx)] and ask whether it is
true under every c-variant. [Fc ↔ (Gc ∨ (∃x)Hx)] is constructed by substituting
c for every free occurrence of x in [Fx ↔ (Gx ∨ (∃x)Hx)]. The bound occurrences
are left unchanged. It is convenient to introduce some notation for talking
about substitution. Let Sb(c/ x)P be the result of substituting c for every
free occurrence of x in the formula P. Then we can formulate a precise
truth rule for universal generalizations as follows:
(9) A universal generalization (∀x )P is true under an interpretation if,
and only if, when we choose some individual constant c that does not
occur in P, Sb( c/ x)P is true under every c-variant of the initial
interpretation.
That Sb(c/ x)P is true under every c -variant of the initial interpretation is
simply another way of saying that Sb(c/ x)P would be true regardless of
what we let the denotation of c be.
Some examples follow of the use of this rule to determine whether
closed formulas of the predicate calculus are true under interpretations.
Consider first the closed formula (∀x)(Fx → Gx) and the interpretation in
which the domain is the set of integers {1,2, 3}, the extension assigned to G
is {1,2}, and the extension assigned to F is {2}. By Rule 9, (∀x)(Fx → Gx) is
true under this interpretation if, and only if, (Fc → Gc) is true under every
c-variant of this interpretation. There are three things in the domain, so
there are three c-variants—we can let the denotation of c be either 1, or 2, or
3. So let us consider each one separately and show that (Fc → Gc) is true
under every one of them. First let the denotation of c be 1. Gc is true under
that interpretation if, and only if,1∈{1,2}. This is the case, so Gc is true. Fc
is true under this c-variant if, and only if,1∈{2}, which is not the case. So Fc
is false. Then, by Rule 7 (Fc → Gc) is true under this c-variant. Next
consider the c-variant that assigns the denotation 2 to c. Gc is true under
that c-variant because 2∈{1,2}. So by Rule 7, (Fc → Gc) is true under the
second c-variant. The third (and last) c-variant assigns 3 to c. Both Fc and
Gc are false under this c-variant, because it is not the case that 3∈{2}, and it
is not the case that 3∈{1,2}. Again by Rule 7, (Fc → Gc) is true. Therefore,
(Fc → Gc) is true under each c-variant of our initial interpretation, and so
by Rule 9, (∀x)(Fx → Gx) is true under that interpretation.
For a more complex example, consider how we can compute the truth
value of the formula
THE SEMANTICS OF THE PREDICATE C ALCULUS 127

[(∀x)Hx ↔ (P → ~(∀x)~Hx)]
under the interpretation that assigns the domain {1,2}, and assigns “true” to
P, and the extension {1} to H. First evaluate the truth value of (∀x)Hx.
(∀x)Hx is true under our interpretation if, and only if, Hc is true under each
c-variant of the interpretation. There are two c-variants—that assigning 1
to c, and that assigning 2 to c. Hc is true under the first c-variant because
1∈{1}, but Hc is false under the second c-variant because it is not the case
that 2∈{1}. Thus Hc is not true under every c-variant, and hence (∀x)Hx is
not true under the initial interpretation. Thus the left side of the biconditional
is false. Next consider (∀x)~Hx. This is true under the interpretation if,
and only if, ~Hc is true under each c-variant of the interpretation. But by
Rule 4, ~Hc is true under an c-variant if, and only if, Hc is false under that
c-variant. We have seen that Hc is true under the first c-variant and false
under the second, so ~Hc is false under the first c-variant and true under
the second. Therefore (∀x)~Hx is not true under the initial interpretation.
P is true, so then by Rule 7, (P → ~(∀x)~Hx) is true under the interpretation.
Hence we have a biconditional the left side of which is false, and the right
side of which is true, so according to Rule 8 the biconditional is false under
the initial interpretation.
It should be mentioned that interpretations will frequently be chosen
which assign sets of integers as domains. This is only because integers are
convenient to work with. These interpretations do not have any privileged
status among interpretations other than their convenience. We could in
principle use any set of objects for the domain of an interpretation.
Next a truth rule for existential generalizations must be constructed.
The rule here is completely analogous to the case of universal generalizations.
We want a closed formula of the form (∃x)Fx to be true if, and only if, there
is something in the domain which is an F; that is, if, and only if, there is at
least one thing in the domain such that, if we choose it as the denotation of
c, Fc will be true. So our rule is the following:
(10) An existential generalization (∃ x)P is true under an interpretation if,
and only if, when we choose some individual constant c that does not
occur in P, Sb(c/ x)P is true under at least one c-variant of the
interpretation.
Consider some examples of the use of Rule 10. Let us calculate the
truth value of (∃x)Bxg under the interpretation that assigns the set of all
past presidents of the United States as the domain, assigns George Washington
as the denotation of g, and assigns to B the set of all ordered pairs of past
presidents such that the first member of each ordered pair was president
before the second member. Here B is the relation “(1) was president before
(2)”. To determine the truth value of (∃x)Bxg under this interpretation we
must first choose an individual constant not occurring in Bxg. We cannot
choose g, because it occurs in Bxg, so let us choose b. Then (∃x)Bxg is true
under this interpretation if, and only if, Bbg is true under at least one
b-variant of this interpretation. But there is no denotation we can assign to
THE SEMANTICS OF THE PREDICATE C ALCULUS 128

b making Bbg true, because such a denotation would have to be a president


before Washington, and there was none. Therefore Bbg is not true under
any b-variant of our interpretation, and so (∃x)Bxg is false under the original
interpretation.
Consider a more complicated example in which one quantifier occurs
within the scope of another quantifier, viz., the closed formula (∀x)(∃y)Lxy
and the interpretation that assigns the domain {1,2, 3} and in which L is the
“less than” relation. The extension assigned to L is {〈1,2〉, 〈1, 3〉, 〈2, 3〉}.
(∀x)(∃y)Lxy is true under this interpretation if, and only if, (∃y)Lcy is true
under every c-variant of this interpretation. There are three c-variants—those
in which we assign either 1, or 2, or 3 to c. But now, (∃y)Lcy is true under a
given c-variant if, and only if, Lcb is true under some b-variant of that
c-variant. For each c-variant there are three b-variants of it—those assigning
1,2, or 3 to b. We can diagram the whole thing as follows:

 b=1 Lcb false 


 
c=1  b=2 Lcb true  (∃y)Lcy true 
  
 b=3 Lcb true  



 b=1 Lcb false  
  
c=2  b=2 Lcb false  (∃y)Lcy true  (∀x)(∃y)Lxy false
  
 b=3 Lcb true  



 b=1 Lcb false  
  
c=3  b=2 Lcb false  (∃y)Lcy false 
 
 b=3 Lcb false 

Lcb is true under the second and third b-variant of the first c-variant,
so (∃y)Lcy is true under the first c-variant. Lcb is true under the third
b-variant of the second c-variant, so (∃y)Lcy is also true under the second
c-variant. But Lcb is not true under any b-variant of the third c-variant, so
(∃y)Lcy is false under the third c-variant. Therefore, it is not true that
(∃y)Lcy is true under every c-variant of the initial interpretation, and so
(∀x)(∃y)Lxy is false under the initial interpretation.
It will often be unnecessary to work through all of the steps of the
truth rules in order to determine whether a closed formula is true under a
given interpretation. If we can see what a closed formula means under an
interpretation we can often see immediately whether it is true. Recall the
above interpretation of (∃x)Bxg. Under that interpretation, (∃x)Bxg means
“There was someone who was president before Washington”. Seeing this,
we know immediately that the closed formula is false. Or consider the
interpretation of the closed formula (∀x)(∃y)Lxy in the preceding example.
Under that interpretation this closed formula means “For every integer
there is another integer such that the first is smaller than the second”, that
is, “For every integer there is a larger integer”. If the domain had been the
THE SEMANTICS OF THE PREDICATE C ALCULUS 129

set of all integers this would have been true, but as the domain is just {1,2,
3} it is false—there is no integer in the domain larger than 3. Explicit appeal
to the truth rules is cumbersome and often unnecessary. However, we
must have such rules for those cases in which the meaning of a closed
formula is so complicated that we cannot tell simply by reflection whether
it is true.

Exercises

Let the universe of discourse be the set {1,2, 3}, and let G mean “is greater
than one”, F mean “is less than three”, H mean “(1) is not equal to (2)”, I
mean “(1) is equal to (2)”, L mean “(1) is less than (2)”, P mean “one is not
equal to two”, and let a denote 1, b denote 2, and c denote 3. This generates
the interpretation that assigns the domain {1,2, 3} to the quantifiers, the
extensions {1,2} to F, {2, 3} to G, {〈1,2〉, 〈1, 3〉, 〈2,1〉, 〈2, 3〉, 〈3,1〉, 〈3,2〉} to H,
{〈1,1〉, 〈2,2〉, 〈3, 3〉} to I, and {〈1,2〉, 〈1, 3〉, 〈2, 3〉} to L, and assigns “true” to P,1
to a,2 to b, and 3 to c. Under this interpretation, express the meaning of
each of the following closed formulas in idiomatic English and determine
whether it is true:
1. Fc 11. (∀x)(∀y)(Ixy → ~Hxy)
2. [(Fa & Fb) & Fc] 12. (∃x)(∀y)(Lxy ∨ Ixy)
3. (∀x)Fx 13. (∀x)(∀y)(Lxy → ~Ixy)
4. (∃x)(∀y)(Fx → Fy) 14. [(∀x)Gx ↔ ((Ga & Gb) & Gc)]
5. (∃x)Gx 15. (∃y)(∀x)(Lxy ∨ Icx)
6. (Iab ∨ ~Iab) 16. (P ↔ (∃x)~Ixx)
7. (Iab ∨ ~Hab) 17. (∀x)(∀y)(∀z)[(Lxy & Lyz) → Lxz]
8. (∀x)(∃y)Ixy 18. (∀x)~Lxx
9. (∃y)(∀x)Ixy 19. (∀x)(∀y)(Lxy → ~Lyx)
10. ((∀x)Ixx & (∀x)~Hxx)

4. Valid Formulas
4.1 Validity and Formal Necessity
What makes it possible to undertake a mathematically precise
investigation of formal necessity in the propositional calculus is the fact that
tautologicity and formal necessity coincide. That is, a formula is a tautology
THE SEMANTICS OF THE PREDICATE C ALCULUS 130

if, and only if, it expresses a formally necessary statement form. Because
tautologicity is a mathematically precise concept, this facilitates the study of
formal necessity.
We can proceed in an analogous way in the predicate calculus. A
closed formula of the predicate calculus is said to be valid if, and only if, it is
true under every interpretation. We write ¤ A to indicate that a formula A
is a valid. It can then be shown that a formula of the predicate calculus is
valid if, and only if, it expresses a formally necessary statement form.
However, the argument for this is more involved than the analogous
argument for the propositional calculus.
We have seen that it is useful to be able to symbolize statements using
a restricted universe of discourse. However, that creates difficulties for
connecting validity and formal necessity, so let us first consider what happens
when we do not use restricted universes of discourse. In other words, let
us assume that the universe of discourse is the universal set. If a closed
formula of the predicate calculus is true under every interpretation, then in
particular it will be true under every interpretation whose domain is the
universal set. Therefore, if this closed formula is what we obtain when we
symbolize a statement (without using a restricted universe of discourse),
then that statement must be true. The statement is made true by its logical
form, so it is formally necessary. We can conclude then that if a closed
formula is valid, then any statement having that form is formally necessary.
Now it must be seen whether the converse is true. Is it true that if a
statement can be symbolized in the predicate calculus, and it is formally
necessary, then it is valid? At first this may seem dubious, because valid
formulas must be true under all interpretations, including those with small
domains like {1,2,3}. We cannot get such a domain by symbolizing a statement
while using the universal set as the universe of discourse. It is apparent
that if a formula A results from symbolizing the form of a formally necessary
statement P, then the formula A must be true under every interpretation
whose domain is the universal set. This is because every such interpretation
must correspond to a statement having the same logical form as P, and
since P is formally necessary, every statement having that logical form
must be true. But this only shows that A must be true under every
interpretation having this one fixed domain. In order to show that A is
valid we must show that it is true under every interpretation having any
nonempty domain. Surprisingly, this can be shown. This turns upon what
is known as the Löwenheim Theorem, which says:
If a formula A of the predicate calculus is true under every interpretation
having some particular infinite domain, then A is true under every
interpretation having any nonempty domain (that is, A is valid).

The universal set is clearly infinite, because among other things it


contains all numbers. Therefore, by the Löwenheim Theorem, if A is true
under every interpretation whose domain is the universal set, then A is
valid. And therefore, if a formula A results from symbolizing the form of a
THE SEMANTICS OF THE PREDICATE C ALCULUS 131

formally necessary statement, then A is valid.


Combining the above two results, we have the following: a formula of
the predicate calculus is valid if, and only if, it expresses a formally
necessary statement form. Thus we can study formal necessity by studying
the mathematically precise concept of validity.

4.2 Using Restricted Universes of Discourse


The preceding assumes, however, that we do not employ restricted
universes of discourse when symbolizing statement forms. If we do, we
can run into a problem that makes the preceding conclusion incorrect. For
instance, suppose we take the universe of discourse to be the set of unicorns,
which is the empty set. Letting R stand for “is red”, we will symbolize the
statement “There is a unicorn that is either red or not red” as (∃x)(Rx ∨
~Rx). The statement “There is a unicorn that is either red or not red” is
false, because there are no unicorns. But the closed formula (∃x)(Rx ∨ ~Rx)
is valid. That is, it is true under every interpretation. This is because
(∃x)(Rx ∨ ~Rx) is true under an interpretation if, and only if, (Rc ∨ ~Rc) is
true under some c-variant of that interpretation. But in fact, (Rc ∨ ~Rc) will
be true no matter what we assign as the denotation of c. So (Rc ∨ ~Rc) will
be true under every c-variant of the interpretation, and (∃x)(Rx ∨ ~Rx) will
be true under the interpretation. Therefore, (∃x)(Rx ∨ ~Rx) is valid, but we
got it by symbolizing a false statement. This difficulty arises from the fact
that interpretations must have nonempty domains, but universes of discourse
can be empty.
We might try avoiding this difficulty by precluding empty universes
of discourse. That will guarantee that if the formula produced by symboling
a statement is valid, then the statement is true. But it will not guarantee
that the statement is necessary. That is because although the universe of
discourse is nonempty, it may not be logically necessary that it is nonempty.
To illustrate, suppose we symbolize the statement “There is a horse that is
either a palomino or not a palomino” as (∃x)(Px ∨ ~Px). Because the latter
formula is valid, it is true in every interpretation, and so in particular it is
true in that interpretation in which the domain is the set of horses and P
expresses “is a palomino”. The difficulty is that it is not a necessary truth
that there are horses, and if there were no horses then the statement “There
is a horse that is either a palomino or not a palomino” would be false
despite the validity of (∃x)(Px ∨ ~Px). So the validity of the formula does
not ensure the necessary truth of the statement.
Although the validity of the formula does not ensure the necessary
truth of a statement symbolized using a restricted universe of discourse, it
does guarantee that if the universe of discourse is nonempty then the statement
is true. So this conditional will be a necessary truth. And if we employ a
universe of discourse consisting of mathematical entities, e.g., the set of all
numbers, then it is a necessary truth that the universe of discourse is
nonempty. That is, it is necessarily true that there are numbers. So in that
case, the problem goes away. But in general, we must be more careful
when using restricted universes of discourse.
THE SEMANTICS OF THE PREDICATE C ALCULUS 132

We have seen that validity provides an analogue for formal necessity


only at the expense of our being unable to use restricted universes of discourse
in symbolizing the forms of statements. This was a very useful shortcut.
There is, however, a way that it can be salvaged. Just as in the propositional
calculus, the concept of validity will be used for assessing the validity of
argument forms by seeing whether their corresponding conditionals are
valid. Now, suppose we have an argument form that we have symbolized
using a restricted universe of discourse. For example, the argument form
Everyone has a father.
Everyone who has a father has a mother.
Therefore, someone has mother.
might be symbolized as
(∀x)Fx
(∀x)(Fx → Mx)
(∃x)Mx
by using the class of people as a restricted universe of discourse. Suppose
then that we form the corresponding conditional
{[(∀x)Fx & (∀x)(Fx → Mx)] → (∃x)Mx}
and ascertain that it is valid. It does not follow from this that the initial
argument is valid. This is because should the universe of discourse be
empty—should there be no people in the world—the corresponding
conditional of the initial argument would be false but its symbolization is
still valid. However, it does follow that another argument, which we obtain
from the initial argument by adding an additional premise, is valid:
There exists at least one person.
Everyone has a father.
Everyone who has a father has a mother.
Therefore, someone has mother.
Here we have just added the premise that the universe of discourse is
nonempty. Why is this new argument valid? The conditional {[(∀x)Fx &
(∀x)(Fx → Mx)] → (∃x)Mx} is valid, so it is true under every interpretation
having a nonempty domain. Therefore, if the class of people is nonempty,
then
{[(∀x)Fx & (∀x)(Fx → Mx)] → (∃x)Mx}
is true under every interpretation whose domain is the class of people.
Therefore, if the class of people is nonempty, then if the two premises of the
original argument are true, the conclusion will be true as well (because
THE SEMANTICS OF THE PREDICATE C ALCULUS 133

these statements correspond to one interpretation of the conditional). But


the new premise we have added to the original argument is the premise
that the class of people is nonempty. Therefore, given all three of these
premises, it follows that the conclusion must be true, and so the new three-
premise argument is valid.
Generally, given an argument symbolized using a restricted universe
of discourse, it does not follow from the validity of the corresponding
conditional that the argument is valid. But it does follow from the validity
of the corresponding conditional that a new argument, which we construct
by adding the additional premise that there is at least one thing in the
domain, is valid. We will usually know that this new premise is true, and
so we can use the argument to conclude that if the original premises are
also true then the conclusion is true. However, if there is some doubt about
whether the universe of discourse is nonempty, then in order to infer the
consequent of the original argument from its premises, we must symbolize
the argument all over again without using a restricted universe of discourse.

4.3 Related Concepts


Related to the concept of validity are three other concepts. We say
that a closed formula is invalid if, and only if, it is not valid. And we say
that a closed formula is consistent (often called satisfiable) if, and only if, its
negation is invalid. A closed formula is inconsistent if, and only if, it is not
consistent. Thus a closed formula is invalid if, and only if, it is false under
at least one interpretation; a closed formula is consistent if, and only if, it is
true under at least one interpretation; and a closed formula is inconsistent
if, and only if, it is false under every interpretation.

4.4 Establishing Validity and Invalidity


In order to show that a closed formula is consistent it suffices to find
one interpretation under which it is true, and in order to show that a closed
formula is invalid it suffices to find one interpretation under which it is
false. But to show that a closed formula is valid you must show that it is
true under every interpretation, and in order to show that a closed formula
is inconsistent you must show that it is false under every interpretation.
Thus, to establish either that a closed formula is valid or that it is inconsistent,
you must give some sort of general argument about all interpretations.
However, to establish that a closed formula is consistent, or that it is invalid,
you need merely give an example of an interpretation under which it is
true, or false, respectively. It must be emphasized that you cannot establish
that a closed formula is valid (or that it is inconsistent) by looking at a
single interpretation, although you can establish that it is invalid (or that it
is consistent) by looking at a single interpretation.
Consider some examples. Suppose we want to show that (∀x)(Fx →
Gx) is invalid. Then we must find an interpretation under which it is false.
Since it is much easier to work with interpretations having small domains,
let us try to find an interpretation under which (∀x)(Fx → Gx) is false which
has a one-element domain. Suppose the domain is {1}. In order to make
THE SEMANTICS OF THE PREDICATE C ALCULUS 134

(∀x)(Fx → Gx) false under some interpretation having this domain, we


must make (Fc → Gc) false under some c-variant of that interpretation. But
there will only be one c-variant, because there is only one object in the
domain. So the denotation assigned to c must be 1. Then in order for (Fc →
Gc) to be false, we must make Fc true and Gc false. Thus,1 must be a
member of the extension of F, but not of G. This means that we must
choose the extension {1} for F and the extension ∅ for G (because if 1 is not
to be in the extension of G, then nothing can be because there is nothing
else in the domain). Thus, (∀x)(Fx → Gx) is false under the interpretation
whose domain is {1} and that assigns the extension {1} to F and ∅ to G.
Therefore, (∀x)(Fx → Gx) is invalid.
This example illustrates that we can use the truth rules to guide us in
searching for interpretations under which formulas have a desired truth
value. However, there is often an easier way of doing it. We might try to
think of meanings we can assign to the relation symbols and individual
constants in a closed formula which will make it symbolize a true statement.
For example, (∀x)(Fx → Gx) means “Every F is a G”. If we want to make
this false we might let F stand for “is a number” and G for “is even”. Then
the formula symbolizes the false statement “Every number is even”. Thus
(∀x)(Fx → Gx) is false under that interpretation whose domain is the set of
numbers, and which assigns the set of numbers as the extension of F and
the set of even numbers as the extension of G.
Next let us show that (∀x)(Fx → Gx) is consistent. To do this we must
find an interpretation under which it is true. Again let us see if we can do it
with the simplest possible domain—a one-element domain, using the domain
{1} again. In order for (∀x)(Fx → Gx) to be true under an interpretation
having this domain, (Fc → Gc) must be true under every c-variant of the
interpretation. But once again there will only be one c-variant—that assigning
1 to c. So when the denotation of c is 1, we must have (Fc → Gc) true.
There are several ways we could do this. We can either have both Fc and
Gc true, or Fc false and Gc true, or both Fc and Gc false. Let us choose the
first alternative. In order to make both Fc and Gc true, the extensions of
both must contain 1, and their extensions cannot contain anything else,
because there is nothing else in the domain. So the extensions must be
simply {1}. Therefore, (∀x)(Fx → Gx) is true under the interpretation whose
domain is {1} and that assigns the extension {1} to both F and G.
Again, this problem might be done more simply by looking for meanings
we can assign to F and G that make the formula express a true statement.
For example, we might consider the interpretation whose domain is the set
of people, and which assigns the set of bachelors as the extension of F and
the set of unmarried people as the extension of G. Then (∀x)(Fx → Gx)
symbolizes the truth statement that all bachelors are unmarried, and so it is
true in that corresponding interpretation.
Consider a more complex example. To show that (∀x)(∃y)(Hxy &
~Hyy) is consistent, we must find an interpretation under which it is true.
Again we can start off by seeking an interpretation having the one-element
domain {1}. Then in order for (∀x)(∃y)(Hxy & ~Hyy) to be true under an
THE SEMANTICS OF THE PREDICATE C ALCULUS 135

interpretation having this domain, (∃y)(Hcy & ~Hyy) must be true under
every c-variant of the interpretation. But there is only one c-variant, that
assigning 1 to c. So (∃y)(Hcy & ~Hyy) must be true under the c-variant of
the interpretation that assigns 1 to c. But (∃y)(Hcy & ~Hyy) is true under
that c-variant if, and only if, (Hcb & ~Hbb) is true under at least one b-variant
of that c-variant. But again, there is only one b-variant, that assigning 1 to
b. So (Hcb & ~Hbb) must be true under some interpretation that assigns 1
to both c and b. But this is impossible, because if c and b have the same
denotation, then Hcb will be true if, and only if, Hbb is true (because it is
the same ordered pair in both cases), and so (Hcb & ~Hbb) cannot be true.
This shows that there cannot be an interpretation having a one-element
domain in which (∀x)(∃y)(Hxy & ~Hyy) is true.
Let us see if we can find an interpretation having a two-element domain,
{1,2}, in which (∀x)(∃y)(Hxy & ~Hyy) is true. This closed formula will be
true under an interpretation having this domain if, and only if, (Hcb &
~Hbb) is true under some b-variant of each c-variant of the interpretation.
There are two c-variants—those assigning 1 and 2 to c—and there are two
b-variants of each c-variant—those assigning 1 and 2 to b. Is there an
extension we can assign to H that will give this result? First consider the
c-variant assigning the denotation 1 to c. Under some b-variant of this
c-variant (Hcb & ~Hbb) must be true. For (Hcb & ~Hbb) to be true we
must have Hcb true and Hbb false. To have Hcb true under some b-variant
of c-variant, there must be some ordered pair in the extension of H whose
first element is 1. The second element must either be 1 or 2, depending
upon which denotation b has. If we allow Hcb to be true under the b-variant
that assigns 1 to b, then 〈1,1〉 will be a member of the extension of H. But
then Hbb will also be true, and (Hcb & ~Hbb) will then be false; so this
b-variant will not do the trick. Suppose then that we let Hcb be true under
the other b-variant—that assigning 2 to b. Then 〈1,2〉 must be in the extension
of H. Also, in order to have Hbb false, 〈2,2〉 must not be in the extension of
H. Thus, by requiring that 〈1,2〉 is in the extension of H, and 〈2,2〉 is not in
the extension of H, we can make (Hcb & ~Hbb) true in some b-variant of
the first c-variant. Now let us see if we can also make it true in the second
c-variant. Here we assign 2 to c. Again, we want to make (Hcb & ~Hbb)
true under some b-variant. Let us see whether we can make it true under
the b-variant that assigns 1 to b. In order to make Hcb true and Hbb false
under that b-variant of this c-variant, we must have 〈2,1〉 in the extension of
H, and we must not have 〈1,1〉 in the extension of H. Therefore, we see that
we can make (∀x)(∃y)(Hxy & ~Hyy) true under an interpretation whose
domain is {1,2} if we can assign an extension to H which contains 〈1,2〉 and
〈2,1〉, but does not contain 〈1,1〉 or 〈2,2〉. We can do this simply by assigning
the extension {〈1,2〉, 〈2,1〉}. Therefore, (∀x)(∃y)(Hxy & ~Hyy) is consistent.
We can also attempt to solve this problem by looking for a restricted
universe of discourse and a meaning for H that will make the formula
express a true statement. If we let the universe of discourse be the set of
positive integers, and we let H mean “(1) is less than (2)”, then (∀x)(∃y)(Hxy
& ~Hyy) means “For every positive integer, there is a positive integer greater
THE SEMANTICS OF THE PREDICATE C ALCULUS 136

than it but not greater than itself”. Thus (∀x)(∃y)(Hxy & ~Hyy) is true
under that interpretation whose domain is the set of positive integers, and
which assigns to H the extension {〈x, y〉  x is less than y}. Another meaning
we might assign to H to make (∀x)(∃y)(Hxy & ~Hyy) true is “(1) is not
equal to (2)”. Then (∀x)(∃y)(Hxy & ~Hyy) means “Given any positive
integer, there is a positive integer that is unequal to it but not unequal to
itself”. Notice that if we let H mean this, and we let our universe of discourse
be {1,2}, we arrive at the same interpretation as we did in the preceding
paragraph merely by considering the truth rules. However, it may be hard
to find such intuitive interpretations for complex formulas. Using the truth
rules for guidance is often a better way to find interpretations making complex
formulas true.
It is generally much easier to work with interpretations having small
domains, so when trying to find an interpretation under which a closed
formula is false, it is generally wise to begin with the smallest possible
domain, a one-element domain. If this does not work, then try a two-element
domain, and so on. Occasionally, however, a closed formula will only be
true under an interpretation having an infinite domain. The closed formula
{(∀x)(∀y)(∀z)[(Rxy & Ryz) → Rxz] & [(∀x)~Rxx & (∀x)(∃y)Rxy]}
can be shown to be false under any interpretation having a finite domain.
But it is consistent, because it can be made true in an interpretation having
an infinite domain. If we let the domain be the set of all positive integers,
and let R be the relation “(1) is a positive integer less than (2)”, this closed
formula is true. So if one cannot find an interpretation having a small
domain under which a closed formula is true, it does not automatically
follow that the closed formula is inconsistent. You might have to try an
infinite domain in order to make the closed formula true.
We have seen that we can show a closed formula to be consistent or
invalid by finding a single interpretation under which it is true, or false,
respectively. But it is more difficult to show that a closed formula is
inconsistent or valid. In the propositional calculus, truth tables provide a
mechanical procedure for determining whether a formula is valid (i.e., a
tautology). Truth tables can get long, but they are still finite structures that
in principle survey all of the possible assignments of truth values to the
atomic parts of the formula. There is no similar mechanical procedure that
can be used for testing validity in the predicate calculus. There are always
infinitely many ways of interpreting the quantifiers, relation symbols, and
individual constants in a formula of the predicate calculus. There is no way
to survey infinitely many interpretations, so there can be nothing like truth
tables for the predicate calculus. Instead, we must give some sort of general
argument to show that the closed formula is either true under all
interpretations (valid), or else false under all interpretations (inconsistent),
depending upon what we are trying to show. Let us look at some examples
of this.
Suppose first that we want to establish the validity of (∀x)(Px → Px).
This closed formula is valid if, and only if, it is true under every interpretation.
THE SEMANTICS OF THE PREDICATE C ALCULUS 137

So consider an arbitrary interpretation. (∀x)(Px → Px) is true under that


interpretation if, and only if, (Pc → Pc) is true under every c-variant of that
interpretation. Just as in the propositional calculus, (Pc → Pc) is true under
every interpretation. A c-variant of an interpretation is an interpretation, so
it follows that (Pc → Pc) is true under every c-variant of any interpretation.
Therefore, (∀x)(Px → Px) must be true under every interpretation; that is,
(∀x)(Px → Px) must be valid.
Next suppose we want to show that (∃y)(Fy & ~Fy) is inconsistent —
that is, it is false under every interpretation. Again, consider an arbitrary
interpretation. (∃y)(Fy & ~Fy) will be true under that interpretation if, and
only if, (Fc & ~Fc) is true under some c-variant of that interpretation. But in
order for (Fc & ~Fc) to be true, Fc would have to be both true and false,
which is impossible. Thus, (Fc & ~Fc) is not true under any c-variant of the
interpretation, and (∃y)(Fy & ~Fy) is false under the interpretation. This is
true for any interpretation we might choose, so (∃y)(Fy & ~Fy) will be false
under every interpretation, and so inconsistent.
To take a more complicated case, let us show that (∃x)(Fx → (∀y)Fy) is
valid. Consider any interpretation. (∃x)(Fx → (∀y)Fy) will be true under
that interpretation if, and only if, (Fc → (∀y)Fy) is true under some c-variant
of that interpretation. (Fc → (∀y)Fy) will be true under a given c-variant as
long as we do not have Fc true and (∀y)Fy false. (∀y)Fy does not contain c,
so it is true under the c-variant if, and only if, it is true under the original
interpretation. Now we have two cases to consider:
Case 1. Suppose (∀y)Fy is true under the original interpretation. Then
(Fc → (∀y)Fy) will be true under every c-variant of the original interpretation
(because it has a true consequent), and so (∃x)(Fx → (∀y)Fy) will be true
under the original interpretation.
Case 2. Suppose (∀y)Fy is false under the original interpretation, then
Fb must be false under some b-variant of the original interpretation; that is,
there must be something we can assign as the denotation of b that will
make Fb false. But now, suppose we consider the c-variant that assigns the
same denotation to c as it does to b. Then Fc will also be false. But then (Fc
→ (∀y)Fy), being a conditional with a false antecedent, will be true under
that c-variant. Thus (∃x)(Fx → (∀y)Fy) will be true under the original
interpretation. In either case, (∃x)(Fx → (∀y)Fy) will be true under the
original interpretation. This is true for every interpretation, so (∃x)(Fx →
(∀y)Fy) must be true under every interpretation—it is valid.
It is often quite difficult to give an argument to show that a closed
formula is valid. Clearly it would be advantageous to have some sort of
general rules to help us along, like the rule of tautological implication in the
propositional calculus. In the next two sections we will be concerned with
finding such rules, and then we will use these rules to construct derivations
in the predicate calculus much like derivations in the propositional calculus.

Exercises
THE SEMANTICS OF THE PREDICATE C ALCULUS 138

A. For each of the following closed formulas, find an interpretation under


which it is true:
1. Fc 5. (∀x)(∃y)Hxy
2. (Fa & ~Fb) 6. (∃y)(∀x)Hxy
3. (∀x)Fx 7. (∀x)(Fx → Gx)
4. (∃x)Fxa 8. (∀x)(Fxx → (∃y)Fxy)

B. Recalling that every formula of the propositional calculus is a closed


formula of the predicate calculus, show that if a formula of the
propositional calculus is truth-functionally consistent, then it is a
consistent formula of the predicate calculus.

5. Implication and Equivalence


In the propositional calculus, the concepts of tautological implication
and tautological equivalence were defined in terms of tautologicity. In the
predicate calculus we can define analogous concepts of implication and
equivalence in terms of the concept of validity. Given two closed formulas
A and B of the predicate calculus, we say that A implies B if, and only if,
the conditional (A → B) is valid. So, for example, (∀x)Fx implies Fa, because
the conditional [(∀x)Fx → Fa] is valid. More generally:
A set of formulas A1 ,…,An implies a formula B if, and only if,
¤ [(A1 &…&An) → B].
We abbreviate “A1,…,An implies B” as “A1 ,…,An ¤ B”. Because all tautologies
are valid, it follows that if a closed formula or set of closed formulas
tautologically implies another closed formula, then it implies the other closed
formula. That is, tautological implication is a special case of implication.
To say that a conditional is true under an interpretation means that if
the interpretation makes the antecedent true it also makes the consequent
true. Thus to say that a conditional is valid means that every interpretation
that makes the antecedent true also makes the consequent true. Thus,
A1 ,…,An ¤ B if, and only if, every interpretation making A1 ,…,An all true
also makes B true.
We define equivalence in an analogous manner. We say that two
closed formulas A and B are equivalent if, and only if, they are true relative
to the same interpretations. This is the same as saying that each formula
implies the other. Equivalently, A and B are equivalent if, and only if, ( A
↔ B) is valid. Thus, for example, (∀x)(Fx → Gx) is equivalent to (∀x)(~Gx
→ ~Fx), because [(∀x)(Fx → Gx) ↔ (∀x)(~Gx → ~Fx)] is valid.
All of the implications and equivalences employed in the propositional
calculus still hold in the predicate calculus. There are also a few new ones
that will be particularly useful. In addition to the tautological implications
I1–I15, we have:
THE SEMANTICS OF THE PREDICATE C ALCULUS 139

I16. (∀ x)A ¤ Sb(c/ x)A


I17. Sb(c /x)A ¤ (∃ x)A
Roughly, I16 says that if A is true of everything, then it is true of c . I17 says
that if A is true of c, then it is true of something.
In addition to the tautological equivalences E1 – E20, we have the
following equivalences:

 (∀ x)A eq. ~(∃x )~A



 (∃ x)A eq. ~(∀x )~A
E21. 
 ~(∀x)A eq. (∃x )~A

 ~(∃x)A eq. (∀x )~A
E22. (∀ x)(A ∨ B) eq. [(∀x)A ∨ B] 
 provided there are no free
E23. (B → (∀ x)A) eq. (∀ x)(B → A) 
 occurrences of x in B.
E24. ((∃ x)A → B) eq. (∀ x)(A → B) 
E25. (∀ x)A eq. (∀y)Sb(y /x)A 
 provided y does not occur in A.
E26. (∃ x)A eq. (∃y )Sb(y /x)A 
E23 and E24 are somewhat surprising, so let us see why they are true.
First consider E23. By E4, (B → (∀x )A) is equivalent to (~B ∨ (∀x)A). By
E12, the latter is equivalent to ((∀ x)A ∨ ~B), and by E22 this is equivalent
to (∀x)(A ∨ ~B). (∀x )(A ∨ ~B) is true under a given interpretation if, and
only if, Sb(c /x)(A ∨ ~B) is true under every c -variant of that interpretation
(where c is some individual constant not occurring in ( A ∨ ~B)). By E12,
Sb(c /x)(A ∨ ~B) is equivalent to Sb(c /x)(~B ∨ A), and then by E4 this is
equivalent to Sb(c/ x)(B → A). Thus (∀ x)(A ∨ ~B) is true under an
interpretation if, and only if, Sb(c /x)(B → A) is true under every c -variant
of that interpretation. But the latter is just the condition under which (∀ x)(B
→ A) is true. Therefore, (B → (∀x)A) is equivalent to (∀x)(B → A).
Similarly, ((∃x)A → B) is equivalent to (~(∃ x)A ∨ B), by E4. By E21
this is equivalent to ((∀ x)~A ∨ B), which by E22 is equivalent to (∀ x)(~A ∨
B). By reasoning as above and using E4, this can be seen to be equivalent
to (∀ x)(A → B).
It should be emphasized that in using E22 – E24, B cannot contain any
free occurrences of x. For example, if B did contain free occurrences of x ,
then ((∀x)A ∨ B) would contain free occurrences of x and so would not
even be a closed formula.
E25 and E26 result from the fact that, for example, (∀x)Fx and (∀y)Fy
mean the same thing. They both say simply that everything is an F. In
terms of the truth rules, they are both true under an interpretation if, and
only if, Fc is true under every c-variant of the interpretation. Thus, it makes
no difference which bound variable we use. However, the restriction that y
THE SEMANTICS OF THE PREDICATE C ALCULUS 140

does not occur in A is essential. Without it we would find ourselves saying,


for example, that (∀x)(∀y)Hxy is equivalent to (∀y)(∀y)Hyy. These closed
formulas, however, are not equivalent. The first says that given any two
things, they stand in the relation H to one another, whereas the second says
that everything stands in the relation H to itself (the first quantifier is a
vacuous quantifier and does not affect the meaning).

Exercises

Prove the following metatheorems. The proofs are similar to the proofs
given in the propositional calculus.
1. Show that if a closed formula is implied by a valid closed formula,
then it is valid.
2. Show that implication is strongly transitive.
3. Show that implication is adjunctive.
4. Using the result of 2, show that if A, B, and C are closed formulas,
and A ¤ B, B ¤ C , and C ¤ A, then A, B, and C are all equivalent.

6. Universal Generalization
A principle of fundamental importance in the predicate calculus is the
principle of universal generalization:
Metatheorem: If A is any formula of the predicate calculus that contains
free occurrences of some variable x but not of any other variable, then
given any individual constant c not occurring in A, if Sb(c/x )A is
valid, then (∀ x)A is valid.
For example, consider the formula (Fx ∨ ~Fx). The closed formula (Fc ∨
~Fc) is valid, because it is a tautology. Therefore, by the above principle,
(∀x)(Fx ∨ ~Fx) is valid. This principle is proven as follows. Suppose
Sb(c /x)A is valid. This means that Sb( c/ x)A is true under every
interpretation. In particular, given any one interpretation, Sb(c/ x)A is not
only true under that interpretation, but also under every other interpretation
exactly like that interpretation except that it assigns something different as
the denotation of c. In other words, given any interpretation, Sb(c /x)A is
true under every c-variant of that interpretation. Therefore, given any
interpretation, (∀x )A is true under that interpretation. Thus, (∀x)A is
valid. ■

We use the principle of universal generalization to establish that


THE SEMANTICS OF THE PREDICATE C ALCULUS 141

universal generalizations are valid when we know that instances of the


scope of the quantifier are valid. This is particularly useful when taken
together with the observation that all tautologies are valid. Given any
tautology, we can construct a valid universal generalization by replacing
individual constants by variables and appending universal quantifiers. For
example, the closed formula
[(Fa ↔ ~Gb) ↔ ~(Fa ↔ Gb)]
is a tautology. From this, by universal generalization, we can conclude that
(∀y)[(Fa ↔ ~Gy) ↔ ~(Fa ↔ Gy)]
is valid. Then by a second application of universal generalization, we can
conclude that
(∀x)(∀y)[(Fx ↔ ~Gy) ↔ ~(Fx ↔ Gy)] is valid.
We can construct an analogous rule of universal generalization for
implications:
Metatheorem: Suppose A is any formula of the predicate calculus that
contains free occurrences of some variable x but not of any other variable,
and suppose that c is some individual constant that does not occur in
A. Then if B1,…,Bn are closed formulas none of which contain c, and
B1 ,…,Bn ¤ Sb(c /x)A, it is also true that B1 ,…,Bn ¤ (∀ x)A.
For example, the set of closed formulas (∀x)(Fx → Gx), (∀x)(Gx → Hx)
implies the closed formula (Fc → Hc). Neither (∀x)(Fx → Gx) nor (∀x)(Gx
→ Hx) contains c. Therefore, this set of closed formulas also implies (∀x)(Fx
→ Hx).
This principle is proven as follows. Suppose B1 ,…,Bn are closed
formulas none of which contain c , and B1,…,Bn ¤ Sb(c /x)A, where A does
not contain c either. By definition, this means that [(B1 & ... & Bn) →
Sb(c /x)A] is valid. As B1,…,Bn are closed formulas, they contain no free
occurrences of x, so
Sb(c /x)[(B1 & ... & Bn) → A]
is the same closed formula as [(B1 & ... & Bn)) → Sb(c /x)A]. Thus
Sb(c /x)[(B1 & ... & Bn) → A]
is valid. Furthermore, as c does not occur in any of B1 ,…,Bn or A, c does
not occur in [(B1 & ... & Bn) → A]. Thus, by the first principle of universal
generalization, (∀ x)[(B1 & ... & Bn) → A] is valid. x does not occur free in
(B1 & ... & Bn), so by E23, (∀x)[(B1 & ... & Bn) → A] is equivalent to [(B1 &
... & Bn) → (∀x)A], and consequently this latter closed formula is also valid.
Therefore, the set of closed formulas B1 ,…,Bn implies (∀ x)A. ■

The principle of universal generalization corresponds to a frequently


used pattern of informal reasoning. Suppose we want to prove that something
THE SEMANTICS OF THE PREDICATE C ALCULUS 142

is true of all objects in a certain domain, say the integers. In order to prove
that it is true of all integers, we might reason as follows. First we pick some
arbitrary integer, and show that it is true of that integer. Then we observe
that our proof that it is true of that integer does not assume anything about
that integer which is not true of all integers, and so the same proof will
work for any other integer as well. Thus we conclude that it is true of every
integer. Observing that the proof does not assume anything about the
integer we chose that is not true of all integers is the same as observing that
the premises of our proof do not refer to that integer—that they do not
contain c. So we are just using universal generalization.

Exercise

Using universal generalization, show that (∀x)(Fx → Gx),(∀x)Fx ¤ (∀x)Gx.


7
Derivations in the
Predicate Calculus

1. Rules of Inference
Derivations in the predicate calculus have essentially the same structure
as derivations in the propositional calculus. In particular, we continue to
employ the inference rules P (premise introduction), I (implication), C
(conditionalization), DN (double negation), and R (reductio ad absurdum). We
have added two implications and six equivalences to the list of implications
and equivalences used by rule I, but the rule is conceptually unchanged. In
addition, we will add one new inference rule. This will be based upon the
principle of universal generalization discussed in chapter six. That rule
was formulated as follows:
Suppose A is any formula of the predicate calculus that contains free
occurrences of some variable x but not of any other variable, and suppose
that c is some individual constant that does not occur in A. Then if
B1 ,…,Bn are closed formulas none of which contain c , and B1,…,Bn ¤
Sb(c /x)A, it is also true that B1,…,Bn ¤ (∀x)A.
Recall that lines of a derivation express implications. They tell us that the
premises of the line imply the conclusion drawn on that line (or if a line has
no premises, they tell us that the conclusion is valid). Accordingly, we can
employ universal generalization to reason as follows:
RULE UG: UNIVERSAL GENERALIZATION: Suppose A is a formula of the
predicate calculus, x is a variable, and A does not contain free occurrences
of any variable other than x , and suppose c is some individual constant
that does not occur in A. Then if Sb(c /x)A appears on some line of a
derivation, and it either has no premises or none of its premises contain
c, we can write (∀ x)A on any later line of the derivation, taking for
premise numbers all the premise numbers of Sb(c /x)A.
We will later add an additional rule of inference to make it easier to construct
derivations in the predicate calculus, but for now we will take derivations
to be defined by the rules P, I, C, DN, R, and UG. That is, a derivation in
the predicate calculus is any sequence of lines that can be constructed in
accordance with these six rules of inference.
Now let us look at some sample derivations without, for the moment,
worrying about the strategies involved in constructing them. Suppose we
want to construct a derivation of (∀x)[Fx → (Fx ∨ Gx)]. We might proceed
as follows:
DERIVATIONS IN THE PREDICATE CALCULUS 144

(1) 1. Fa P
(1) 2. (Fa ∨ Ga) (I3), 1
3. [Fa → (Fa ∨ Ga)] C, 1, 2
4. (∀x)[Fx → (Fx ∨ Gx)] UG, 3
To take another example, suppose we want to derive (∀x)(∃y)Fxy from
(∃y)(∀x)Fxy. We might proceed as follows:

(1) 1. (∃y)(∀x)Fxy P
(2) 2. (∀x)Fxa P
(2) 3. Fba (I16), 2
(2) 4. (∃y)Fby (I17), 3
(2) 5. (∀x)(∃y)Fxy UG, 4
6. [(∀x)Fxa → (∀x)(∃y)Fxy] C, 2, 5
7. (∀y)[(∀x)Fxy → (∀x)(∃y)Fxy] UG, 6
8. [(∃y)(∀x)Fxy → (∀x)(∃y)Fxy] (E24), 7
(1) 9. (∀x)(∃y)Fxy (I9), 1, 8
Suppose next we want to derive (∃y)Gy from (∀x)(Fx → Gx) and
(∃z)Fz. We might proceed as follows:

(1) 1. (∀x)(Fx → Gx)P


(2) 2. (∃z)Fz P
(3) 3. Fa P
(1) 4. (Fa → Ga) (I16), 1
(1, 3) 5. Ga (I9), 3, 4
(1, 3) 6. (∃y)Gy (I17), 5
(1) 7. [Fa → (∃y)Gy] C, 3, 6
(1) 8. (∀z)[Fz → (∃y)Gy] UG, 7
(1) 9. [(∃z)Fz → (∃y)Gy] (E24), 8
(1, 2) 10. (∃y)Gy (I9), 2, 9

Exercises

In each of the following purported derivations, check the lines for errors,
and indicate the numbers of the lines on which errors occur. An error is
said to occur in a line if the line cannot be introduced in accordance with
the rules of inference. This means that in checking a line for errors, you
only look at that line to see whether it has the necessary relation to the
previous lines, and treat the previous lines as if they were all correct. If a
line is incorrect, try to find a way of correcting it (perhaps by adding additional
lines).
DERIVATIONS IN THE PREDICATE CALCULUS 145

1. (1) 1. (∀x)(Fx ∨ Gx) P


(2) 2. (∃x)(∀y)(Fx → Gy) P
(3) 3. (∃x)~Gx P
(2) 4. (∀y)(Fa → Gy) (I17), 2
(2) 5. (Fa → Gb) (I16), 4
(3) 6. ~Gb P
(2, 3) 7. ~Fa (I10), 5, 6
(1) 8. (Fa ∨ Ga) (I16), 1
(1, 2, 3) 9. Ga (I11), 7, 8
(1, 2, 3) 10. (∀x)Gx UG, 9
(1, 2) 11. [(∃x)~Gx → (∀x)Gx] C, 3, 10
(1, 2) 12. [~(∀x)Gx → (∀x)Gx] (E21), 11
(1, 2) 13. [~~(∀x)Gx ∨ (∀x)Gx] (E4), 12
(1, 2) 14. [(∀x)Gx ∨ (∀x)Gx] (E1), 13
(1, 2) 15. (∀x)Gx (E18), 14

2. (1) 1. (∃x)[Gx → (∀y)Hxy] P


(2) 2. [Ga → (∀y)Hay] P
(3) 3. Ga P
(2, 3) 4. (∀y)Hay (I9), 2 ,3
(2, 3) 5. Hab (I16), 4
(2) 6. (Ga → Hab) C, 3, 5
(2) 7. (∃y)(Gy → Hyb) (I17), 6
(2) 8. (∀x)(∃y)(Gy → Hyx) UG, 7
9. {[Ga → (∀y)Hay] → (∀x)(∃y)(Gy → Hyx)} C, 2, 8
10. {(∃x)[Gx → (∀y)Hxy] → (∀x)(∃y)(Gy → Hyx) (I17), 9
(1) 11. (∀x)(∃y)(Gy → Hyx) (I9), 1, 10

2. Strategies
Just as in the propositional calculus, there are strategies that can guide
one in the construction of derivations. The strategies for conjunctions,
disjunctions, conditionals, and biconditionals are the same as before. The
strategy for negations must be augmented by the observation that, by E21,
~(∀x)P is equivalent to (∃ x)~P, and ~(∃x)P is equivalent to (∀x )~P. To
these we add strategies for universal generalizations and existential
generalizations.

2.1 Universal Generalizations


When reasoning forward, we may derive conclusions from a universal
generalization, and when reasoning backward we may derive new interests
DERIVATIONS IN THE PREDICATE CALCULUS 146

from interests in universal generalizations. We have separate strategies for


each.

Forward reasoning
Typically, when reasoning forward from a universal generalization
we use I16 to obtain one or more instances of the generalization, and then
continue our reasoning from the instances. This is called universal instantiation.
For example, suppose suppose we want to derive (Fa & Ga) from (∀x)Fx.
We might proceed as follows:

(1) 1. (∀x)Fx P
(1) 2. Fa (I16), 1
(1) 3. Fb (I16), 1
(1) 4. (Fa & Fb) (I14), 1

Note that we can use universal instantiation to infer as many instances of


the universal generalization as may be useful. The only difficult part of this
strategy is figuring out which instances will be useful, and even that is
usually not very difficult.

Backward reasoning
Backward reasoning consists of adopting interest in other conclusions
from which we could infer the formula we are trying to derive. If our
objective is to derive a universal generalization, we will usually do this by
employing UG. For instance, suppose we want to derive (∀x)(~Gx → ~Fx)
from (∀x)(Fx → Gx). Then we construct the following annotated derivation:

(1) 1. (∀x)(Fx → Gx) P


(1) 1. (∀ x)(~Gx → ~Fx)
(1) 2. (~Ga → ~Fa) for 1 by (UG)
(1) 2. (Fa → Ga) (I16), 1
(1) 3. (~Ga → ~Fa) (E16), 2 this discharges interest 2
(1) 4. (∀x)(~Gx → ~Fx) UG, 3 this discharges interest 1

The general strategy is to adopt interest in an instance of the generalization


we are trying to derive. In order to use universal generalization later, the
individual constant employed in the instance must be a “new” constant not
occurring elsewhere in the formula or in any premises of the formula. So
given an interest in a formula (∀x )A, we choose an individual constant c
that does not occur in A or in any of the premises, and adopt interest in
Sb(c /x)A. If we are able to derive the latter, we can then derive (∀ x)A by
universal generalization.
DERIVATIONS IN THE PREDICATE CALCULUS 147

Exercises

For each of the following, construct an annotated derivation of the closed


formula below the line from the closed formulas above the line. Keep the
strategies in mind.

1. (∀x)[Fx → (Hx & ~Gx)]


(Ga → ~Fa)

2. (∀x)Fx
(∀x)(Fx → Gx)
(∀x)Gx

2.2 Existential Generalizations


We also have distinct strategies for reasoning forward or backward
with existential generalizations.

Backward reasoning
To derive an existential generalization from something else, we typically
use I17, deriving the existential generalizatrrion from an instance of it. For
example, if we want to derive (∃x)Fx from (∀x)Fx, we might proceed as
follows:

(1) 1. (∀x)Fx P
(1) 1. (∃ x)Fx
(2) 2. Fa for 1 by (I17)
(1) 2. Fa (I16), 1 this discharges interest 2
(1) 3. (∃x)Fx (I17), 2 this discharges interest 1

Forward reasoning
Our most complex strategy concerns how to reason forward from an
existential generalization. Before considering the general strategy, let us
look at an example of it. To derive (∃x)Gx from (∃x)(Fx & Gx). We might
proceed as follows:

(1) 1. (∃x)(Fx & Gx) P


(1) 1. (∃ x)Gx
(2) 2. (Fa & Ga) P
(2) 2. (∃ x)Gx for 1, by existential instantiation
(2) 3. Ga for 2 by (I17)
(2) 3. Ga (I2), 2 this discharges interest 3
(2) 4. (∃x)Gx (I17), 3 this discharges interest 2
DERIVATIONS IN THE PREDICATE CALCULUS 148

5. [(Fa & Ga) → (∃x)Gx] C, 2, 4


6. (∀x)[(Fx & Gx) → (∃x)Gx] UG, 5
7. [(∃x)(Fx & Gx) → (∃x)Gx] (E24), 6
(1) 8. (∃x)Gx (I9), 1, 7 this discharges interest 1

This strategy is called existential instantiation. In order to derive something


from (∃x)(Fx & Gx), we took as a new premise (Fa & Ga), which is Sb(a/x)(Fx
& Gx). Then we proceeded to derive our desired conclusion (∃x)Gx from
this new premise. Next we conditionalized, obtaining

[(Fa & Ga) → (∃x)Gx]

Then we used UG, to obtain (∀x)[(Fx & Gx) → (∃x)Gx]. Because x does not
occur free in (∃x)Gx, we were able to use E24 to convert the universal
generalization into [(∃x)(Fx & Gx) → (∃x)Gx], and finally, by modus ponens,
we inferred our desired conclusion from the original premises.
Generalizing this strategy, suppose we want to derive some closed
formula B from the premise (∃ x)A together with some other premises (whose
numbers are indicated in the following derivation by the dots in the left-hand
column). We choose an individual constant c, that does not occur in B or A
or any of the premises, and proceed as follows:

(1) 1. ... P 

(2) 2. ... P  premises other

.  than (∃ x)A
. 
. 
(n) n. (∃ x)A P
(1,...,n) 2. B for 1 by existential instantiation
(n + 1) n + 1. Sb(c /x)A P
. 
. 
.  derivation of

 B from Sb(c /x)A

(n + 1, . . . ) k. B  this discharges interest 2
... k + 1. [Sb(c /x)A → B] C, (n + 1), k
... k + 2. (∀ x)(A → B) UG, k + 1
... k + 3. [(∃ x)A → B] (E24), k + 2
(n, . . . ) k + 4. B (I9), n, k + 3
Notice that the second and third derivation on page 144 both use existential
instantiation.
The intuitive rationale for existential instantiation is the following.
Suppose we know that (∃x)Fx is true, i.e., that something is an F. From this
we want to derive a conclusion Q. We might begin our argument by
saying, “We know that something is an F; call it ‘c’”. This is the same as
DERIVATIONS IN THE PREDICATE CALCULUS 149

taking Fc as a new premise. If we can then derive Q from Fc without


assuming anything about c, then Q will follow from Fc regardless of what c
is, and so Q will follow from the mere fact that there is something that is an
F. This reasoning corresponds to the last three steps of existential instantiation.
Notice that in order for us to be able to use UG on line (k + 2), c
cannot occur in A or B or any of the premises. This corresponds in the
intuitive reasoning pattern to requiring that we have not assumed anything
about c . It is important to keep this restriction in mind, because it is a very
common mistake to overlook it.

Exercises

For each of the following, construct an annotated derivation of the closed


formula below the line from the closed formulas above the line. Keep the
strategies in mind.

1. Fa
(∃x)(Fx ∨ Gx)

2. (∃x)Fx
(∃x)(Fx ∨ Gx)

3. (∃x)(Fx ∨ Gx)
~(∃x)Fx
(∃x)Gx

2.3 A General Strategy


These strategies combine to give us one basic strategy for derivations
in the predicate calculus: eliminate quantifiers from the premises using
existential and universal instantiation, eliminate quantifiers from the interests
using existential and universal generalization, then proceed as in the
propositional calculus to derive the conclusion without quantifiers, and
finally replace quantifiers using existential and universal generalization.
Let us look at some examples of this.
To illustrate, suppose we want to derive (∃x)(∀z)(∃y)Fxyz from
(∀x)(∃y)(∀z)Fxyz. We might proceed as follows:

(1) 1. (∀x)(∃y)(∀z)Fxyz P
(1) 1. (∃ x)(∀ z)(∃y)Fxyz
(1) 2. (∃y)(∀z)Fayz (I16), 1
DERIVATIONS IN THE PREDICATE CALCULUS 150

(3) 3. (∀z)Fabz P (beginning existential instantiation)


(3) 2. (∃ x)(∀ z)(∃y)Fxyz for 1, by existential instantiation
(3) 3. (∀ z)(∃ y)Fayz for 2 by (I17)
(3) 4. (∃ y)Fayc for 3 by UG
(3) 5. Fabc for 4 by (I17)
(3) 4. Fabc (I16), 3 this discharges interest 5
(3) 5. (∃y)Fayc (I17), 4 this discharges interest 4
(3) 6. (∀z)(∃y)Fayz UG, 5 this discharges interest 3
(3) 7. (∃x)(∀z)(∃y)Fxyz (I17), 6 this discharges interest 2
8. [(∀z)Fabz → (∃x)(∀z)(∃y)Fxyz] C, 3, 7
9. (∀y)[(∀z)Fayz → (∃x)(∀z)(∃y)Fxyz] UG, 8
10. [(∃y)(∀z)Fayz → (∃x)(∀z)(∃y)Fxyz] (E24), 9
(1) 11. (∃x)(∀z)(∃y)Fxyz (I9), 2, 10 (completing
existential instantiation) this discharges interest 1

Here we began by eliminating quantifiers in conclusions 1 – 4 and interests


1 – 5. This gave us our conclusion without quantifiers. Then in steps 5 – 11
we replaced the quantifiers to give us our conclusion with the desired
quantifiers.
To take a more difficult example, suppose we want to derive (∀x)(Fx
→ ~Px) from the set of premises (∀x)[Gx → (∀y)(Fy → Hxy)], (∀x)[Gx →
(∀z)(Pz → ~Hxz)], and (∃x)Gx:

(1) 1. (∀x)[Gx → (∀y)(Fy → Hxy)] P


(2) 2. (∀x)[Gx → (∀z)(Pz → ~Hxz)] P
(3) 3. (∃x)Gx P
(1,2,3) 1. (∀ x)(Fx → ~Px)
(4) 4. Ga P (beginning existential instantiation)
(1,2,4) 2. (∀ x)(Fx → ~Px) for 1, by existential instantiation
(1,2,4) 3. (Fb → ~Pb) for 2 by UG
(1) 5. [Ga → (∀y)(Fy → Hay)] (I16), 1
(2) 6. [Ga → (∀z)(Pz → ~Haz)] (I16), 2
(1, 4) 7. (∀y)(Fy → Hay) (I9), 4, 5
(2, 4) 8. (∀z)(Pz → ~Haz) (I9), 4, 6
(1, 4) 9. (Fb → Hab) (I16), 7
(2, 4) 10. (Pb → ~Hab) (I16), 8
(2, 4) 11. (~~Pb → ~Hab) DN, 10
(2, 4) 12. (Hab → ~Pb) (E16), 11
(1, 2, 4) 13. (Fb → ~Pb) (I13), 9, 12 this discharges interest 3
(1, 2, 4) 14. (∀x)(Fx → ~Px) UG, 13 this discharges interest 2
(1, 2) 15. [Ga → (∀x)(Fx → ~Px)] C, 4, 14
(1, 2) 16. (∀x)[Gx → (∀x)(Fx → ~Px)] UG, 15
DERIVATIONS IN THE PREDICATE CALCULUS 151

(1, 2) 17. [(∃x)Gx → (∀x)Fx → ~Px)] (E24), 16


(1, 2, 3) 18. (∀x)(Fx → ~Px) (I9), 3, 17 (completing
existential instantiation) this discharges interest 1

In this derivation, conclusions 1 – 10 and interests 1 – 3 involve eliminating


the quantifiers from the premises. Steps 11 – 13 involve getting the desired
conclusion without its quantifier. Steps 14 – 18 are concerned with putting
the quantifier on by UG, and completing the existential instantiation used
to eliminate the quantifier from the third premise.
It should be emphasized that UG and I16 and I17 can only be used
with respect to an entire line—never just part of a line of a derivation. If for
example, we had a derivation in which the closed formula [Fa → (∃y)Gy]
appeared, we could not obtain [(∀x)Fx → (∃y)Gy] by UG, nor could we
obtain [(∃x)Fx → (∃y)Gy] by I17. Similarly, if [(∀x)Fx → (∃y)Gy] appeared
on a line of derivation, we could not use I16 to obtain [Fa → (∃y)Gy]. These
restrictions on I16 and I17 are restrictions on Rule I in general. Remember
that we can never use Rule I on just part of a line. And a similar restriction
applies to Rule UG.
Clearly it will generally be much easier to show that a closed formula
is valid by constructing a derivation of it than by giving some sort of general
argument involving truth under all interpretations. But, just as in the case
of derivations in the propositional calculus, the construction of derivations
is not a purely mechanical matter. It will often require considerable ingenuity
to find the appropriate derivation. Constructing derivations is not materially
different from proving theorems in mathematics. In the light of this, one
might desire some purely mechanical means, analogous to truth tables, for
evaluating the validity or invalidity of closed formulas. Unfortunately, it
can be proven that there is no such mechanical technique, so we will have
to be content with using our ingenuity and constructing derivations.

3. A Shortcut Rule for Existential Instantiation


When using the strategy of existential instantiation, the last four steps
are always the same. We employ rule C, UG, E24, and I9. It becomes
tedious to mechanically write out these four steps each time we use the
strategy, so it is convenient to introduce a shortcut rule (analogous to DN
and R) that allows us to compress those four steps into a single line. For
example, in the preceding derivation, this will allow us to move directly
from line 14 to line 18, just rewriting the premise numbers:
RULE EI: EXISTENTIAL INSTANTIATION If a derivation contains a formula of
the form (∃x)A on some line i and on another line j it contains a premise
of the form Sb(c/x )A where c does not occur in A, then if a formula B
occurs on any later line and (1) c does not occur in B, (2) j is one of the
premise numbers of B, and (3) c does not occur in any of the premises
DERIVATIONS IN THE PREDICATE CALCULUS 152

of B except for j, then we can write B on any later line of the derivation,
replacing the premise number j by the premise numbers of line i.

The condition that c not occur in B or any of its premises other than Sb( c /x)A
is required for us to be able to use UG in inferring (∀ x)(A → B) from
[Sb(c /x)A → B].
Using EI, we can shorten the preceding derivation as follows:

(1) 1. (∀x)[Gx → (∀y)(Fy → Hxy)] P


(2) 2. (∀x)[Gx → (∀z)(Pz → ~Hxz)] P
(3) 3. (∃x)Gx P
(1,2,3) 1. (∀ x)(Fx → ~Px)
(4) 4. Ga P (beginning existential instantiation)
(1,2,4) 2. (∀ x)(Fx → ~Px) for 1by EI
(1,2,4) 3. (Fb → ~Pb) for 2 by UG
(1) 5. [Ga → (∀y)(Fy → Hay)] (I16), 1
(2) 6. [Ga → (∀z)(Pz → ~Haz)] (I16), 2
(1, 4) 7. (∀y)(Fy → Hay) (I9), 4, 5
(2, 4) 8. (∀z)(Pz → ~Haz) (I9), 4, 6
(1, 4) 9. (Fb → Hab) (I16), 7
(2, 4) 10. (Pb → ~Hab) (I16), 8
(2, 4) 11. (~~Pb → ~Hab) DN, 10
(2, 4) 12. (Hab → ~Pb) (E16), 11
(1, 2, 4) 13. (Fb → ~Pb) (I13), 9, 12 this discharges interest 2
(1, 2, 4) 14. (∀x)(Fx → ~Px) UG, 13 this discharges interest 3
(1, 2, 3) 15. (∀x)(Fx → ~Px) EI, 14 (completing
existential instantiation) this discharges interest 1

Exercises

For each of the following, construct an annotated derivation of the closed


formula below the line from the closed formulas above the line. Keep the
strategies in mind.

1. (∀x)(Px → ~Px)
~Pa

2. (∀x)[Fx → (Hx → Gx)]


(∀x)Fx
[(Ha → Ga) & ~(~Gb & Hb)]
DERIVATIONS IN THE PREDICATE CALCULUS 153

3. (∀x)[Px ↔ (Hx & ~Px)] (hint: try reductio)


(∀x)~Hx

4. [(∃x)Fx → (∀y)Gy]
(∀x)(∀y)(Fx → Gy)

5. (∀x)(Fx → Gx)
~(∀x)Gx
(∃y)~Fy

6. (∀x)[Fx → (Gx → Hx)]


[(∀x)(Fx → Gx) → (∀x)(Fx → Hx)]

7. (∀x)(∃y)Rxy
(∀x)(∀y)(Rxy → Ryx)
(∀x)(∀y)(∀z)[(Rxy & Ryz) → Rxz]
(∀x)Rxx

B. Show that each of the following closed formulas is valid by constructing


an annotated derivation of it; that is, a derivation whose last line has no
premise numbers:
1. [(∀x)Fx → (∃x)Fx]
2. (∃x)(Fx → (∀y)Fy) (hint: use reductio ad absurdum)
3. [(∀x)(∀y)(Rxy → ~Ryx) → ~(∃x)Rxx]
4. ~(∃x)(∀y)(Rxy ↔ ~Ryy)
5. ~(∀x)[(Fx ∨ ~Fx) → ~(Fx ∨ ~Fx)]
6. {(∃x)(Fx ∨ Gx) ↔ [(∃x)Fx ∨ (∃x)Gx]}
7. {(∀x)(Fx & Gx) ↔ [(∀x)Fx & (∀x)Gx]}
8. {(∀x)(Fx → Gx) → [(∀x)Fx → (∀x)Gx]}
9. [(P → (∃x)Fx) ↔ (∃x)(P → Fx)]
10. [((∀x)Fx → P) ↔ (∃x)(Fx → P)]
C. Symbolize each of the following arguments in the predicate calculus,
and construct an annotated derivation of its conclusion from its premises.
1. Every teacher has some good students. Some students are also
teachers. Thus some students have some good students. [T: “(1) teaches
DERIVATIONS IN THE PREDICATE CALCULUS 154

(2)”, G: “is a good student”]


2. Caesar was warned (by someone) to beware the ides of March. If he
had been warned by someone reliable, he would have been careful, but
he was not careful. Therefore, he was warned by someone unreliable.
[W: “(1) warned (2) to beware the ides of March”; P: “is a person”; C:
“is careful”; R: “is reliable”; c: Caesar]
3. If someone were to steal the royal crown, he would not be able to
sell it. If a person stole the royal crown and could not sell it, he would
melt it down. If someone melts down the royal crown, he will discover
that it is hollow, and know that the prince is a crook. Therefore, if
someone steals the royal crown, then someone will know that the prince
is a crook. [P: “is a person”; S: “(1) steals (2)”; c: the royal crown; A:
“(1) is able to sell (2)”; M: “(1) melts down (2)”; D: “(1) discovers (2) is
hollow”; K: “(1) knows that (2) is a crook”; p: the prince]
4. Given any positive integer, there is a greater positive integer.
Therefore, given any positive integer, there are two positive integers
such that the second is greater than the first, and the first is greater
than the given integer. [P: “is a positive integer”; L: “(1) is larger than
(2)”]
5. Some people in England are older than anybody in Scotland. But
some people in Siberia are older than anybody in England. If one
person is older than a second person, and the second is older than a
third, then the first person is older than the third. Therefore, some
people in Siberia are older than anybody in Scotland. [E: “is a person
in England”; S: “is a person in Scotland”; O: “(1) is older than (2)”; I: “is
a person in Siberia”]

D. Try your hand at the following derivations. These are famously difficult
derivations taken from the literature on automated theorem proving.1 Some
of these are hard enough that even professional logicians and mathematicians
may require several hours to construct a derivation.

1. Pelletier’s problem 26

[(∃x)Px ↔ (∃y)Qy]
(∀x)(∀y)[(Px & Qy) → (Rx ↔ Sy)]
[(∀x)(Px → Rx) ↔ (∀y)(Qy → Sy)]

1
Most of these are from the list compiled by Jeff Pelletier, “Seventy-five
problems for testing automatic theorem provers”, Journal of Automated Reasoning 2 (1986),
191-216.
DERIVATIONS IN THE PREDICATE CALCULUS 155

2. Pelletier’s problem 29

[(∃x)Fx & (∃y)Gx]


{[(∀x)(Fx → Hx) & (∀y)(Gy → Jy)] ↔ (∀z)(∀w)[(Fz & Gw) → (Hz & Jw)]}

3. Pelletier’s problem 33

– no premises –
{(∀x)[(Pa & (Px → Pb))→ Pc]↔ (∀x)[(~Pa ∨ (Px ∨ Pc)) &(~Pa ∨ (~Pb∨ Pc))]}

4. Pelletier’s problem 34 (Andrew’s Challenge)

– no premises –
{[(∃x)(∀y)[(Px ↔ Py)↔ ((∃z)Qz ↔ (∀w)Qw)]↔ [(∃u)(∀v)(Qu↔ Qv)↔ ((∃x)Px ↔ (∀y)Py)]}

5. Pelletier’s problem 37

(∀z)(∃w)(∀x)(∃y)[(Pxy → Pyw) & (Pyw → (∃u)Quw)]


(∀x)(∀z)[~Pxz → (∃v)Qvz]
[(∃y)(∃z)Qyz → (∀x)Rxx]
(∀x)(∃y)Rxy

6. Pelletier’s problem 42

A set is “circular” if it is a member of another set which in turn is a


member of the original. Pelletier observes that, intuitively, all sets
are non-circular. The problem is to show that there is no set of
noncircular sets.

– no premises –
~(∃y)(∀x)[Fxy ↔ ~(∃z)(Fxz & Fzx)]

7. Pelletier’s problem 43

Define set equality (Q) as having exactly the same members. Prove
set equality is symmetric.

(∀x)(∀y)[Qxy ↔ (∀z)(Fzx ↔ Fzy)]


(∀x)(∀y)(Qxy ↔ Qyx)
DERIVATIONS IN THE PREDICATE CALCULUS 156

8. Pelletier’s problem 45

(∀x){(Fx & (∀y)[(Gy & Hxy) → Jxy]) → (∀y)[(Gy & Hxy) → Ky]}
~(∃y)(Ly & Ky)
(∃x){[Fx & (∀y)(Hxy → Ly)] & (∀y)[(Gy & Hxy) → Jxy]}
(∃x)[Fx & ~(∃y)(Gy & Hxy)]

9. Pelletier’s problem 46

(∀x){(Fx & (∀y)[(Fy & Hyx) → Gy]) → Gx}


[(∃x)(Fx & ~Gx) → (∃x){(Fx & ~Gx) & (∀y)[(Fy & ~Gy) → Jxy]}]
(∀x)(∀y)[(Fx & Fy) & Hxy) → ~Jyx]
(∀x(Fx → Gx)

10. Pelletier’s problem 47. This is known as the Schubert Steamroller


problem. This is a slightly whimsical symbolization of the following:

Wolves, foxes, birds, caterpillars, and snails are animals, and there
are some of each of them. Also, there are some grains, and grains
are plants. Every animal either likes to eat all plants or all animals
much smaller than itself that like to eat some plants. Caterpillars
and snails are much smaller than birds, which are much smaller
than foxes, which in turn are much smaller than wolves. Wolves do
not like to eat foxes or grains, while birds like to eat caterpillars but
not snails. Caterpillars and snails like to eat some plants. Therefore,
there is an animal that likes to eat a grain-eating animal.

(∀x)(Wx → Ax) (∀x)(∀y)[(Cx & By) → Mxy]


(∀x)(Fx → Ax) (∀x)(∀y)[(Sx & By) → Mxy]
(∀x)(Bx → Ax) (∀x)(∀y)[(Bx & Fy) → Mxy]
(∀x)(Cx → Ax) (∀x)(∀y)[(Fx & Wy) → Mxy]
(∀x)(Sx → Ax) (∀x)(∀y)[(Wx & Fy) → ~Exy]
(∃w)Ww (∀x)(∀y)[(Wx & Gy) → ~Exy]
(∃f)Ff (∀x)(∀y)[(Bx & Cy) → Exy]
(∃b)Bb (∀x)(∀y)[(Bx & Sy) → ~Exy]
(∃c)Cc (∀x)[Cx → (∃y)(Py & Exy)]
(∃s)Ss (∀x)[Sx → (∃y)(Py & Exy)]
(∃g)Gg (∀x)(Gx → Px)
(∀x)[Ax → [(∀w)(Pw → Exw) → (∀y)((Ay & (Myx & (∃z)(Pz & Eyz))) → Exy)]]
(∃x)(∃y)[(Ax & Ay) & (∃z)[Exy & (Gz & Eyz)]]

11. The “unintuitive problem”. This problem is of interest because when


presented to most logicians or mathematicians, their immediate reaction is
“That cannot be right”.
DERIVATIONS IN THE PREDICATE CALCULUS 157

(∀x)(∀y)(∀z)[(Pxy & Pyz) → Pxz]


(∀x)(∀y)(∀z)[(Qxy & Qyz) → Qxz]
(∀x)(∀y)(Qxy → Qyx)
(∀x)(∀y)(~Pxy → Qxy)
~Pab
Qcd

12. This is a problem in group theory taken from Chang and Lee, Symbolic
Logic and Mechanical Theorem Proving, Academic Press, 1973. Pxyz symbolizes
“x·y = z”.

(∀x)Pxex
(∀x)Pexx
(∀x)(∀y)(∀z)(∀u)(∀v)(∀w){[Pxyu & (Pyzv & Puzw)] → Pxvw}
(∀x)(∀y)(∀z)(∀u)(∀v)(∀w){[Pxyu & (Pyzv & Pxvw)] → Puzww}
(∀x)Pxxe
Pabc
Pbac

You might also like