Page 108 - Foundations of Cognitive Psychology : Core Readings
P. 108

Minds, Brains, and Programs  109

                 Still, there are several reasons why AI must have seemed—and to many
               people perhaps still does seem—in some way to reproduce and thereby explain
               mental phenomena, and I believe we will not succeed in removing these illu-
               sions until we have fully exposed the reasons that give rise to them.
                 First, and perhaps most important, is a confusion about the notion of ‘‘infor-
               mation processing’’: many people in cognitive science believe that the human
               brain, with its mind, does something called ‘‘information processing,’’ and
               analogously the computer with its program does information processing; but
               fires and rainstorms, on the other hand, don’t do information processing at all.
               Thus, though the computer can simulate the formal features of any process
               whatever, it stands in a special relation to the mind and brain because when the
               computer is properly programmed, ideally with the same program as the brain,
               the information processing is identical in the two cases, and this information
               processing is really the essence of the mental. But the trouble with this argu-
               ment is that it rests on an ambiguity in the notion of ‘‘information.’’ In the sense
               in which people ‘‘process information’’ when they reflect, say, on problems in
               arithmetic or when they read and answer questions about stories, the pro-
               grammed computer does not do ‘‘information processing.’’ Rather, what it does
               is manipulate formal symbols. The fact that the programmer and the interpreter
               of the computer output use the symbols to stand for objects in the world is to-
               tally beyond the scope of the computer. The computer, to repeat, has a syntax
               but no semantics. Thus, if you type into the computer ‘‘2 plus 2 equals?’’ it will
               type out ‘‘4.’’ But it has no idea that ‘‘4’’ means 4 or that it means anything at
               all. And the point is not that it lacks some second-order information about the
               interpretation of its first-order symbols, but rather that its first-order symbols
               don’t have any interpretations as far as the computer is concerned. All the
               computer has is more symbols. The introduction of the notion of ‘‘information
               processing’’ therefore produces a dilemma: either we construe the notion of
               ‘‘information processing’’ in such a way that it implies intentionality as part of
               the process or we don’t. If the former, then the programmed computer does not
               do information processing, it only manipulates formal symbols. If the latter,
               then, though the computer does information processing, it is only doing so in
               the sense in which adding machines, typewriters, stomachs, thermostats, rain-
               storms, and hurricanes do information processing; namely, they have a level of
               description at which we can describe them as taking information in at one end,
               transforming it, and producing information as output. But in this case it is up
               to outside observers to interpret the input and output as information in the
               ordinary sense. And no similarity is established between the computer and the
               brain in terms of any similarity of information processing.
                 Second, in much of AI there is a residual behaviorism or operationalism.
               Since appropriately programmed computers can have input–output patterns
               similar to those of human beings, we are tempted to postulate mental states in
               the computer similar to human mental states. But once we see that it is both
               conceptually and empirically possible for a system to have human capacities in
               some realm without having any intentionality at all, we should be able to
               overcome this impulse. My desk adding machine has calculating capacities, but
               no intentionality, and in this paper I have tried to show that a system could
   103   104   105   106   107   108   109   110   111   112   113