Professional Documents
Culture Documents
64
CHAPTER 5
Cognition as Computing
This chapter takes a closer look at the fundamental claims of computationalism. The
discussion divides into four sections. The first explicates the central theses of
computationalism and analyzes how they are related to one another. The second looks into the
concept of computation as defined in terms of the Turing machine. The third examines the
difference between the classical and connectionist models of computationalism and
investigates how this difference bears on the theses of computationalism. And the fourth
examines the kind of intelligence assumed by proponents of computationalism and likewise
investigates how it bears on the same theses.
Herbert Simon and Craig Kaplan (1989, 2) define cognitive science as the study of
intelligence and its computational processes in humans (and animals), in computers, and in
the abstract. This definition identifies the levels on which a computationalist investigation of
the nature of intelligence is to be carried out; namely, on the abstract level, on the human (and
animal) level, and on the machine level. Based on these levels, the central claims of
computationalism can be classified in terms of a general thesis, which concerns the abstract
level of intelligence, and two sub-theses, which concern the levels of humans and machines.
65
say, if it is really true that cognition is a species of computing then any entity, not just humans
or machines, considered to be intelligent must be an entity whose intelligence is a species of
computing or whose intelligence is a computational process.
The general thesis concerns two features; namely, the feature of being cognitive,
which we can refer to as cognitivity, and the feature of being computational, which we can
refer to as computationality. The relation between these two features is not that they are
identical nor that computationality falls under cognitivity but that cognitivity falls under
computationality. Meaning to say, while it is not necessary that all computational systems are
cognitive, it is necessary that all cognitive systems are computational. Furthermore, the points
of reference for these two features are not the same. Machines are the point of reference for
the computational feature is machines, while humans are the point of reference for the
cognitive feature. Humans are given as cognitive systems and machines are judged to be
cognitive or not based on their similarities to humans; while machines are given as
computational systems and humans are judged to be computational or not based on their
similarities to humans. It is not the case that machines are given as computational and
cognitive and then humans are judged to be computational and cognitive, or vice-versa. These
considerations explain why the thesis of machine intelligence does not claim that all
computing machines are cognitive, only those capable of simulating human cognitive process;
but the thesis of human computationality claims that all humans capable of cognition are
computational in terms of their cognitive processes.
Now, if the basis for saying that computing machines are cognitive is that they are
capable of simulating human cognitive processes, what about the claim that human cognition
is necessarily computational, what is its basis? Definitely not that the fact that humans are
capable of simulating the computing processes of a machine, for to begin with these
computing processes of machines are products of human cognition. Meaning, these machines
are human artifacts and their computational processes are designed by humans. The mere fact
that humans perform computations, as they do when doing mathematical calculations, does
not necessarily mean that their cognitive processes are computational. What it only entails is
that the human mind is capable of performing computations, which may be one of the many
types of processes that the human mind can possibly also perform.
66
There are definitely many questions that will be raised as a result of our observations.
But we need to begin with the basics before we can appropriately deal with these questions. In
what follows, we shall be dealing with the following concerns: what it really means to say that
something is computational and what it really means to say that something is cognitive in the
context of computationalism.
What does it really mean to say that thinking is a kind of computing? But first what does it
really mean for anything to be computing? Computing is generally the process of
implementing a computation, which is also called an algorithm. To compute is simply to
implement a computation. What then is a computation? Computation can generally be defined
as a step-by-step effective procedure for getting a desired result. Harnish (2002, 125) puts it as
follows: The essence of an algorithm (or effective procedure) is that it is a series of steps for
doing something that is guaranteed to get a result. More precisely, it is a finite sequence of
well-defined steps, each of which takes only a finite amount of memory and time to complete,
and which comes to an end on any finite input. Computing is not limited to solving
mathematical problems or functions, as algorithms are general in that they apply to any
problem. Thus, there can be an algorithm or an effective procedure for cooking eggs, for
instance, or for washing the laundry.
As Crane (1995, 88) explains:
In this regard, thinking as computing means nothing but that thinking is a process of
carrying out certain computations or effective procedures. But more importantly, it also means
that thinking process can be spelled out in terms of a well-defined series of steps. But what
kind of computations should there be such that when carried out constitutes thinking? For
obviously not all sorts of implementing computations constitute thinking. As we noted earlier
there can be an effective procedure, and hence a computation, for cooking eggs but the
process of cooking eggs does not constitute thinking.
Roger Schank (1984, 68) describes what researchers in the discipline of artificial
intelligence intend to accomplish in the following way: AI researchers are still trying to
determine the processes that comprise intelligence so that they can begin to develop computer
algorithms that correspond to human algorithms for understanding. This statement, to begin
with, gives credence to our previous observation that the general thesis of computationalism
concerns intelligence in the abstract which is then applied to humans. What is problematic
here is that the abstract thesis is not derived from the case of human intelligence but from the
case of machine intelligence. We shall pursue this issue later on. But for our present purposes
67
this statement of Schank speaks of human algorithms for understanding, which can also be
referred to as cognitive algorithms. Accordingly, these cognitive algorithms then constitute
the type of computations such that their implementation constitutes thinking. But what these
cognitive algorithms are is precisely the project of artificial intelligence. Be it as it may, the
underlying assumption is that thinking operates via the implementation of certain
computations.
The notion of computation has initially been defined in terms of the Turing machine. The
Turing machine does not refer to a particular type of machine that one can buy and use. It is
an abstract specification of any possible computing machine. It specifies the basic operations
that a physical system must be capable of performing in order to qualify as a computing
machine. These basic operations involve two operations; namely, receiving an input and
executing a command or instruction to produce an output upon reading an output. For these
operations to be possible there must, however, be the following: (1) a storage system, (2) a
reading and printing device, (3) a set of symbols, (4) a translation system, and (5) a set of
instructions or commands.
The storage system can be anything so long as it is divided into certain portions such
that a bit of data or information can symbolically be stored in it. In the case of Turing, he
conceives of the storage system (corresponding to what is presently called memory) as a
tape of infinite length that is divided into squares. In each square, a symbol can be written in
case the square is empty or does not yet contain a symbol. If the square already contains a
symbol, such symbol can be erased and either be left empty or a symbol (either a new one or
the same as the one erased) be written anew. What are written and stored in the tape are
symbols, which are also called representations for they are intended to represent certain data.
These symbols are finite in number but they can be combined in unlimited ways. These
symbols serve as the language of the machine for its operations (corresponding to the 0s and
1s of modern computers). And of course the machine does not receive inputs or information
in the form of the symbols that it uses. The machine translates the inputs that it receives into
its own language, so to speak. And this requires that the machine has a translation system.
This translation, which is the symbolization of the input data, allows the machine to operate
on a very general level, very much in the same way that the use of symbols in logic allows us
to speak of reasoning in a very general manner. The generality of the language of the machine
enables the machine to receive a wide variety of data and process them in multifarious ways.
For the symbols to be written, stored, erased, and overwritten, there must be a
scanning and printing device. Since the storage system is a tape consisting of squares, this
device must be capable of moving the tape from left to right and vice-versa as it reads, stores,
erases, and overwrites symbols in the squares. And of course, whatever the device does is in
line with the machines set of instructions or commands, called by Turing as the machine
table. These commands are in the conditional form. For instance, if it reads 0 in square A,
it should move to square B and write 1 or if it reads 10 in square X, it should move to
square Y and overwrite the symbol already written there with the symbol 1010. Now, as the
machine is performing a particular task, Turing describes the machine as being in a particular
68
internal state. There is nothing subjective or mysterious about these states being internal;
their being internal only means that they refer to the physical states of the machine on the
level of performing its tasks and not on the level of the physical composition of the machine
(the physical states of the physical composition of the machine, in this lights, would be the
machines external states). To use the viewpoint of functionalism, these internal states are the
functional states of the machine, which take place on the level of the functional organization
of the machine, in contrast to its lower-level physical states, which take place on the level of
its material components.
This machine was originally conceived by Alan Turing for purposes of determining
the computability of any given mathematical function; that is to say, whether there is a
mathematical procedure that can be used to find the value of a given mathematical function,
or what comes to the same, whether there is a computation or an algorithm to find the value of
a given mathematical function. If a given mathematical function is computable then there
must be a computation or a mathematical procedure to find the value of such function. Before
we proceed with our discussion, it may be asked, but what is a mathematical function? Here,
Gottlob Freges discussion of the nature of a function is helpful (see Frege 1960).
69
classes of problems in mathematics are not computable or soluble, an example of which is the
halting problem.) (see Penrose 1994, 28-29). And what results from Turings machine is a
definition of what computation is or what computability consists in. Accordingly, a
computation is anything that can be run by a Turing machine. This definition, which is widely
accepted among mathematicians, has come to be known as the Church Thesis or sometimes as
the Church-Turing Thesis [for it is said that Alonzo Church has independently arrived at the
same conclusions as Turings (see Crane 1995, 99, Penrose 1994, 20-21)), and though another
logician, Emil Post, has done the same even earlier than Church (see Penrose 1994, 20)].
Turings genius and amazing discoveries, however, do not stop here. In the course of
developing his concept of the Turing machine, he then realizes that all possible Turing
machines could be run by a single Turing machine, or more precisely, all the operations of
various Turing machines could be performed by a single machine. If the operations of all
Turing machines can be reduced to the same basic operations, then a single machine will
suffice to run the operations of all other Turing machines. What is simply needed is for the
table machines (the programs) of the other Turing machines be inputted in the tape (memory)
of this single machine. This single Turing machine is called the Universal Turing Machine. If
in principle all the individual Turing machines can be put in this universal machine, then the
storage capacity of this universal machine must be unlimited. This partly explains why the
tape of the Turing machine is infinite in length.
The Universal Turing Machine is in fact the theoretical forerunner or model of the
modern-day, general-purpose, digital computer. Think, for instance, of the individual
machines that enable us to do mathematical calculations (the calculator), to view DVD
movies (the DVD player), to view televisions shows (the TV set), to hear music (the radio,
CD player, and MP3 player), to organize our activities (the electronic organizer), to
communicate with other people (the telephone and cell phone), to write and print documents
(the electronic typewriter), and play games (the play station). Each of these machines is an
instantiation of a particular Turing machine, an input-output device with memory (the tape)
and program (the table machine). But all of them can be put in one single machine, as it is
already done in our present computers. The only difference between the Universal Turing
Machine and our present computers is the storage capacitythe formers is unlimited while
the latters is still limited though rapidly expanding as time passes by.
If computation is defined in terms of being run in a Turing machine and the computer
is the approximate embodiment of the Universal Turing Machine, then a computation can also
be defined in terms of the actions of the computer, as Roger Penrose (1994, 17) does: What
is a computation? In short, one can simply understand that term to denote the activity of an
ordinary general-purpose computer. To be more precise, we must make take this in a suitably
idealized sense: a computation is the action of a Turing machine. Consequently, any
computational system is a Turing machine; and the human mind, as a computational system
that has enormous storage capacity and the capacity to run multiple types of operations is
(more precisely, an instantiation of) a Universal Turing Machine, and hence is a computer.
Furthermore, as the computer differs from the Universal Turing Machine as a matter of
degree, then it follows that the difference between the human mind and the computer is
likewise a matter of degree. Human intelligence and computer intelligence are of the same
70
nature; they are just different types of the same process: the process of computing (see
Pylyshn 1989, 52).
At the present, it is usual to distinguish between two types of computationalism: the classical
or symbolic and the connectionist. The computationalism that we have discussed thus far is of
the classical type. It is called classical because it is the type of computationalism that has been
existing prior to the advent of the connectionist type. It is, however, called symbolic because
it regards computational process as a process performed over symbols or representations
either in the case of humans or in the case of machines. A classic pronouncement to this effect
comes from Simon and Kaplan (1989, 40): Computer, and (in our view) human beings are
symbols systems. They achieve their intelligence by symbolizing external and internal
situations and events and by manipulating those symbols. They all use about the same
symbol-manipulating process. Pylyshn (1989, 57) makes the same explanation: The
important thing is that, according to the classical view, certain kinds of systems, including
both minds and computers, operate on representations that take the form of symbolic codes.
Basically, therefore, from a symbolic computationalist point of view, computation is
essentially a symbol-manipulating process.
71
that the intelligent machine is a symbol system, thus making a specific architectural
assertion about the nature of intelligent systems.
Fodor was greatly influenced by the idea of the famous linguist-philosopher Noam
Chomsky that we are born with the same linguistic categories that enable us to learn various
natural languages. This view is Chomskys alternative to the claim of the behaviorist B. F.
Skinner that the learning of natural languages is a matter of conditioning or association
between stimuli and responses, which Chomsky has shown to be mistaken. In the hands of
Fodor, these innate linguistic categories have become the language of thought. The language
of thought may be compared to the internal language of computers, or better yet, of our cell
phones. As Fodor (1979, 67) explains: On this view, what happens when a person
understands a sentence must be a translation process basically analogous to what happens
when a machine understands (viz., compiles) a sentence in its programming language.
Notice that we can easily change the natural language of our cell phones, say from English to
Filipino or German, for these languages are just programmed into the internal language or
the language of the main program of our cell phones. As such, the rules governing the
computational states of our cell phones are not the grammatical rules of the natural languages,
but the rules of their internal language. In the same way, our mental states do not follow the
rules of our natural languageswhich, in a way, are just programmed into our minds, but the
rules of the language of thought.
Fodor advances three arguments for his language-of-though hypothesis, namely, (1)
the productivity of thoughts, (2) the systematicity of thoughts, and (3) the systematicity of
reasoning. The productivity of thoughts refers to the capacity of the human mind/brain to
produce many new thoughts from a number of old thoughts by combining these old thoughts
in novel ways. Systematicity of thoughts, on the other hand, refers to the capacity of the
human mind/brain to understand new thoughts built on already-understood old thoughts. For
instance, anyone who understands John loves Mary will also understand Mary loves
John, or anyone who understands the statement A small red ball is in a large blue box will
also understand the statement A large blue ball is in a small red box. And systematicity of
reasoning refers to the capacity of the human mind/brain to recognize inferences that are of
the same logical structure. For instance, anyone who can infer the statement It is sunny
from the statement It is sunny and warm and humid (or can infer P from P and Q and R) can
also infer the statement It is sunny from the statement It is sunny and warm (or can also
infer P from P and Q). The idea is that since the features of productivity and systematicity
involving ordinary language or linguistic representations are made possible by the logical or
formal structure of language, the productivity and systematicity involving the mind or mental
72
representations must also be made possible by a system that has a structure similar to the
logical or formal structure of language.
To begin with, the units, also called nodes, are the basic elements of a connectionist
system called a network or net. These units receive, process, and send information, and they
are interconnected thereby forming a network or net. Based on their functions, the units are
of three kinds: the input units, which receive information from outside the network and to be
processed in the network; the output units, which carry the resulting processed information;
and the hidden units, which are the units in between the input units and output units. The
hidden units, which may come in various layers, receive information from the input units,
process these data, and then pass these processed data to the output units. The direction of the
flow of information, or activations, may be forward, that is from the input units to the hidden
units and then to the output units; or it may be recurrent, where the activations, after passing
from the input units to the hidden units, pass through several layers of hidden units in a back
and forth manner first before finally sending the processed information to the output units.
The flow of information among the units is made possible by the connections among
the units. These connections are said to have weights or strengths, which affect the amount of
information flowing through them. The weight of a connection is called a connection weight.
Each unit is said to have a maximum amount of data that it can receive, called its threshold
level. It is when the information received by a unit exceeds its threshold level that it is
activated to pass information to other units. The amount of information that will be received
by a unit will therefore be a combination of the strength of the connection through which the
information passes through and the amount of information given off by the sending unit.
Now where does computation come in? The strengths of the connections among the
units can be adjusted, and so given a certain input to a network, the output it will give off can
be manipulated. This process of adjusting the connections among the units of a network is
called the process of teaching the network, and it is this process that is governed by a
computation. A computation here specifies how much adjustments should be done to the
connections of a network such that given a certain input the network will give off a certain
output. On the other hand, these networks are called neural for it is believed, though
arguably, that the connectionist model of computing is close to how the human brain works.
The units correspond to the human brains neurons and the connections correspond to its
synapses.
73
Based on our general accounts of the two types of computationalism, the main
difference between symbolic computationalism and connectionism is that in symbolic
computationalism a physical system processes an input to generate a desired output by
manipulating symbols following a certain program, while in connectionism a physical system
processes an input to generate a desired output by adjusting the connections among the units
in networks following a computation or a learning program. There are other various ways of
expressing this difference. One is that computation in the classical type is serial for it is a step
by step process, while that in the connectionist type is parallel for the interactions among the
untis in networks take place simultaneously. This accounts for why connectionism is also
sometimes referred to as PDPparallel distributed processing. Another is that computation
in the classical type is described as symbolic for computations are carried out through and
over symbols, while that of the connectionist type is described as sub-symbolic for the
computations are carried without the use of symbols but by means of the adjustments of the
connection weights. But if in connectionism, computation does not make use of symbols, how
are information represented? There are two possible ways: localized representation, where
one unit is made to represent one item of information; and distributed representation, where
one item of information is represented by the patterns of connections among the units.
In response to the connectionist model, Fodor and Pylyshn (Fodors close ally), in a
joint publication (Connectionism and Cognitive Architecture: A Critical Analysis, 1993),
have launched criticisms against this model. The main contention is that only a symbolic type
of computationalism can account for the features of productivity and systematicity of thoughts
and reasoning undeniably at work in mental processes. Fodor and Pylyshn further claims that
at best connectionism is just a theory concerning the implementation of symbolic
computationalismof how symbolic computationalism works. For these reasons, they do not
regard connectionism as a rival to classical computationalism but simply as a sub-theory of
classical computationalism. In any case, the difference between classical and connectionist
computationalism does not touch the core of the claims of computationalism. For whether
computation is symbol-manipulation or interaction among the units in neural networks, it is
still computation. But since the classical model is, relatively speaking, well-established type
of computationalism, we shall assume this type of computationalism in discussing
computationalism throughout the remaining part of this book.
74
Although no really satisfactory intentional definition of intelligence has been
proposed, we are ordinarily willing to judge when intelligence is being exhibited by
our fellow human beings. We say that people are behaving intelligently when they
choose courses of action that are relevant to achieving their goals, when they reply
coherently and appropriately to questions that are put to them, when they solve
problems of lesser or greater difficulty, or when they create or design something useful
or beautiful or novel. We apply a single term, intelligence, to this diverse set of
activities because we expect that a common set of underlying processes is implicated
in performing all of them.
Schank (1984, 62) states that AI is an attempt to build intelligent machines, not people. The
idea being that AI does not attempt to reproduce in machines the whole of human mentality.
Intelligent machines may be as intelligent as humans, but they are not humans; and this is
precisely because being human is not just being intelligent. Schank makes this clear when he
distinguishes among three levels of human understanding or intelligenceSchank regards
intelligence as inextricably linked to understanding. From the lowest to highest, these three
levels of understanding are the levels of making sense, cognitive understanding, and complete
empathy. Schank then argues that among these levels of human understanding the ones that
machines can ever get to achieve are the levels of making sense and cognitive understanding.
The level of complete empathy is exclusive to humans. Let us elaborate on this point by
Schank.
First, on the level of making sense, understanding simply means knowing what is
happening in a given situation; and hence this level of understanding only requires simple
recognition of the terms used and the actions performed in such situation (Schank 1984, 45).
Second, on the level of cognitive understanding, understanding means being able to do some
of the following: (1) learn or change as a result of its experiences; (2) relate present
experiences to past experiences intelligently, i.e., by making useful comparisons and picking
out important and interesting relationships; (3) formulate new information for itselfcome to
its own conclusions based on experience; and (40 explain itselfsay why it made the
connections it did, and what thought process was involved in reaching a conclusion (Schank
1984, 47-48). With computers, experiences would mean inputs. Lastly, on the level of
complete sympathy, understanding means identifying with or being able to relate with the
experiences and feelings of another. The assumption here is that the one who understands and
75
the one being understood share some memories and experiences, or that ones memories and
experiences resemble those of the other (Schank 1984, 45). To further illustrate the differences
of these levels, let us take a look at the following example of Schank (1984, 57-58):
INPUT:
Did I tell you about what happened with Mary last night? We were at this party
drinking and talking about nothing in particular and all of a sudden we werent talking
anymore and I noticed her face was right near mine. She moved closer and we kissed.
OUTPUTMAKING SENSE
You and Mary were at a party. You were close to each other. You kissed each other.
You didnt talk while kissing.
OUTPUTCOGNITIVE UNDERSTANDING:
Mary must like you. From what youve told me about other women you know, she is
more assertive than they are. She must have felt good being with you at the party.
OUTPUTCOMPLETE EMPATHY:
Thats like what happened to Cindy and me after the party. She came up to me and
asked if I could give her a lift, and while we were in the coatroom she threw her arms
around me and started kissing me.
In this example, one understands the situation on the level of making sense because
one is able to recognize what is happening in the situation. On the level of cognitive
understanding, one understands it because one is able to relate present experiences to past
experiences and formulate new information based on this relation. On the level of complete
sympathy, one understands it because one is able to identify or relate with the experience of
the person in the situation due to ones similar experience in the past. According to Schank
(1984, 58), while a computer that is appropriately programmed can understand the situation
up to the level of cognitive understanding, no computer is going to understand this story at
the level of COMPLETE SYMPATHY for the simple reason that no computer is ever going to
go to parties and get kissed and hugged by people. Actually, it is conceivable that a machine
run by a computer, say an android, to go to a party and get kissed and hugged by people. But
what this machine would not able to know is what it is like to go to a party, get kissed, and
hugged by people. It may have such experiences as inputs, but such inputs would lack the
same qualities that go with such experiences if experienced by humans. And this explains why
no computer would be able to identify with a human having such experiences in a certain
qualitative way. In any case, Schank (1984, 45) concludes that:
Computers are not likely to have the same feelings as people, or to have the same
desires and goals. They will not be people. The complete empathy level of
understanding seems to be out of reach of the computer for the simple reason that the
computer is not a person.
There are two consequences of this considerationthat it is not really the goal of
cognitive science and AI to explain the mind in its entirety in computational terms. The first is
76
that computationalism is never really meant to be an exhaustive theory of the mind. The
second is that the usual criticisms leveled against computationalism seem to be misplaced.
These criticisms refer to those arguments that show that computationalism fails to account for
or leave out in their explanation the other important features of mentality, such as qualia or
consciousness and intentionality. These criticisms are built on the understanding (or
misunderstanding) that computationalism and the disciplines that espouse it would also like to
explain in computational terms human feelings in addition to human intelligence. Consider,
for instance, the following remarks by Roger Penrose (1989):
Suppose also that the machines have been carefully programmed and fed with great
quantities of data of an appropriate kind. The manufacturers are claiming that the
devices actually think. Perhaps they are also claiming them to be genuinely intelligent.
Or they may go further and make the suggestion that the devices actually feelpain,
happiness, compassion, pride, etc. (p. 5-6)
One of the claims of AI is that it provides a route towards some sort of understanding
of mental qualities, such as happiness, pain, hunger. (p. 14)
The supporters of strong AI would claim that whenever the algorithm were run it
would, in itself: experience feelings; have a consciousness; be a mind. (p. 180)
That being the case, it could, however, be argued that intelligence, strictly speaking, is
a conscious mental activity. One cannot simply talk of intelligence and not talk of
consciousness at the same time. This, actually, is the position taken by Penrose when he
argues that intelligence cannot take place without awareness. Penrose (1994, 38-39) regards
the terms intelligence, understanding, and awareness as related in the following way:
(a) intelligence requires understanding and (b) understanding requires awareness. In
his view (Penrose 1994, 39), consciousness has a passive aspect, which refers to awareness,
and an active aspect, which refers to the feeling of free will. Intelligence or the activity of
thinking is a necessarily conscious phenomenon, and thus to say that something can be
intelligent without being conscious is to misuse or to deviate from the original meaning of the
word intelligence. Consequently, attributing intelligence to machines necessarily implies
attributing consciousness to machines as well. And this perhaps explains why he attributes to
strong AI the claim that machines can be constructed such that it will not only be capable of
thinking but of feeling as well.
77
are not really capable of genuine understanding, though their output behaviors seem to
suggest that they are.
Contrast the above view with how AI scientists regard intelligence. Schank (1984, 51,
39)) writes that When we ask What is intelligence? we are really only asking What does an
entity, human or machine, have to do or say for us to call it intelligent? and We really have
no standards for measuring intelligence other than by observing the behavior of something we
suspect is intelligent. A while ago, we discussed Schanks levels of understanding, notice that
these levels are distinguished in terms of the kind of output one exhibits if presented with a
situation or a storyas in the case of the example we discussed. Observe again how Simon
and Kaplan (1989, 1) define cognitive scienceas the study of intelligence and intelligent
systems, with particular reference to intelligent behavior as computation; and intelligence
people are behaving intelligently when they choose courses of action that are relevant to
achieving their goals, when they reply coherently and appropriately to questions that are put
to them, when they solve problems of lesser or greater difficulty, or when they create or
design something useful or beautiful or novel (1989, 1). Observe also how Simon and
Newell (1995, 96-97) express their Physical Symbol System Hypothesis and define what an
intelligent action is: A physical symbol system has the necessary and sufficient means for
general intelligent action.By general intelligent action we wish to indicate the same scope
of intelligence as we see in human action: that in any real situation behavior appropriate to the
ends of the system and adaptive to the demands of the environment can occur, within some
limits of speed and complexity.
We can see that these AI scientists define intelligence not in terms of consciousness
but in terms of functionality. Thus, we can distinguish between two contrasting views
concerning the nature of intelligence. We can call these views the Functional View and the
Common-sense View. The former is the one adhered to by computationalists, cognitive
scientists, and AI scientists; while the latter is the one adhered to by the likes of Penrose and
Searle. Intelligence undoubtedly has a functional and a conscious aspect. Its functional aspect
has to do with the type of activities that it can perform or tasks it can accomplish as well as
the type of processes that it undergoes to perform certain activities or to accomplish certain
tasks. This aspect involves all the behavioral activities that Schank and Simon and Kaplan talk
about concerning an intelligent action or behavior like solving problems and replying
appropriately to certain questions. Its conscious aspect, on the other hand, has to do with the
fact that the processes it undergoes to perform certain activities or accomplish certain tasks is
accompanied by consciousness. And by consciousness, we basically mean awareness, which
always comes in a certain qualityqualia or the phenomenal property, and which sometimes
involves awareness of what mental states mean or represent in the worldintentionality or
the intentional property.
Now when we distinguish between the functional view and the common-sense view,
we do not mean to distinguish views that deny the existence of either the functional or the
conscious properties of intelligence. I think it will be extremely difficult to support a view that
denies either the functional properties or the conscious properties of intelligence. The
78
distinction only concerns which between these two types of properties is primary or definitive
of the concept of intelligence. Accordingly, the functional view does not deny the reality of
the conscious properties of intelligence but claims that it is the functional properties of
intelligence that are important to the concept of intelligence. On the other hand, the common-
sense view does not deny the functional properties of intelligence but it claims that it is the
conscious properties that are important to the concept of intelligence. Another way of putting
this is that for the functional view it is the functional properties that primarily define the
concept of intelligence, while for the common-sense view it is the conscious properties that
primarily define such concept. Thus, the question is not whether or not intelligence is possible
without consciousness but whether it is the functional properties of intelligence that primarily
defines the concept of intelligence or its conscious properties.
Some believe that the Turing test is not necessary to the claims of AI and cognitive
science. Chalmers (n.d., 8-9), for instance, writes:
Instead, the central property of computation on which I will focus is one that we have
already noted; the fact that a computation provides an abstract specification of the
causal organization of a system. Causal organization is the nexus between computation
79
and cognition. If cognitive systems have their mental properties in virtue of their
causal organization, and if the causal organization can be specified computationally,
then the thesis of computation sufficiency is established. Similarly, if it is the causal
organization of a system that is primarily relevant in the explanation of behavior, then
the thesis of computational explanation will be established.
But there is a sense in which strong AI needs the Turing test. And this is to determine
what sort of computations or programs will produce intelligence in machines. The problem,
however, is that this test only works for systems already established to be computational in
nature, such as computers. Since it does not establish the computational nature of intelligence,
only the intelligence of a computational system, it does not apply to human intelligence. In
fact, in the original test, human intelligence is the given and is not what is in question.
4. Concluding Remarks
Since the computational theory of mind focuses on human intelligence and not on human
mentality in its entirety, it can be said that its conclusions are only intended to apply to the
functional aspect of human intelligencethat aspect of our intelligence which we can share
with machines. In this light, the arguments purporting to show the implausibility of this theory
by showing that its conclusions are precisely limited only to the functional aspect of the mind
seem misplaced. However, even if we grant this narrow application of the claims of this
theory, there is still reason to doubt such claims. For the fact that our functional intelligence is
computable or lends itself to a computational simulation does not necessarily establish that it
is computational. That is to say, while our functional intelligence can be simulated
computationally such that its behavioral manifestations can be replicated by a machine, such
behavioral manifestations need not be the result of a computational process or of the process
of implementing a set of computations as performed by our mind. We shall further pursue this
line of argumentation in the next chapter.
80
81