Peirce, Searle, and the Chinese Room Argument

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

Send inquiries to: Steven Ravett Brown, 714 Ingleside Drive, Columbia, MO 65201, USA; srbrown@ravett.com. Affiliated with the Philosophy Department, University of Oregon, Eugene, OR.


 

Abstract

 

Whether human thinking can be formalized and whether machines can think in a human sense are questions that have been addressed since the Renaissance. I will employ arguments from both a modern critic, John Searle, and from one present at the inception of the field, Charles Peirce, and another inductive argument, all of which conclude that digital computers cannot achieve human-like understanding. Searle approaches the problem from the standpoint of traditional analytic philosophy. Peirce would have radically disagreed with Searle’s analysis, but he ultimately arrives at the same conclusion. Given this diversity of arguments against the Artificial Intelligence (AI) project, it would seem its ultimate goal is futile, despite the computer’s amazing achievements. However, I will show that those arguments themselves imply a direction for AI research which seems fruitful and which is in fact being pursued, although it is not in the mainstream of that field.

 

 


 

There are three classes of meta-analyses concerning the problems of whether human thinking can be formalized and/or whether machines can think in a human sense. The first class, the “mysterian”, denies that either of these is possible because of mysterious properties of mind or consciousness which we cannot analyze or duplicate physically. The second class denies that these are possible on the grounds that our conceptual repertoire is insufficient, perhaps intrinsically insufficient, to the task, i.e., that the questions may be answerable, but not by human beings, or at least not by humans at our present level. This claim, at its strongest, relates to conceptualizing in a deep sense, involving something like the Kantian categories, and maintains that our fundamental human categories of understanding are inadequate to the task, perhaps irredeemably so; at its weakest, it asserts that our theories and understanding of the world are at present inadequate in some fundamental sense. The third class is basically that of present-day scientific materialism, which argues that the lack of such solutions is a result of our lack of knowledge and theory, and that this lack can be remedied through further research and theoretical insight. In considering some of the arguments for the inability of the digital computer to realize mind, then, I will immediately dismiss the first and second positions, not on the grounds of agreement or disagreement, but rather on the grounds that, given either of those positions, it is clearly the case that the answer is negative to both questions. It is the third position which makes for perhaps the most heated arguments, and it is this position that I will assume in this essay.

 We are currently attempting to emulate or realize mind through our artifacts, with some limited success. Engineers and computer scientists are pursuing this goal with no regard for the arguments of philosophers; yet those arguments, while largely irrelevant to the details of applications involved, might have great impact, if clearly enough formulated, on the general directions of research. Dreyfus, for example, in What Computers Can’t Do (Dreyfus, 1972), sparked numerous debates within the AI community. I will concern myself here with two other critics of AI. John Searle is a modern critic of the AI position; his arguments are more analytically oriented than those of Dreyfus and thus perhaps more relevant to the reductive approach underlying most of AI. Charles Peirce was present at the inception of the field of AI, and his analyses of mind and of logic are in part directed toward its conceptual underpinnings. In addition, I will present an inductive argument against finite-state devices and AI toward the end of this essay. I will suggest that consideration of the relationships between these various positions might update this dispute and indicate a direction which might prove fruitful in resolving the issue as to what type of machine might ultimately demonstrate “mindedness”.

 

Searle’s argument (Searle, 1990), as we will see, is based on an analysis of language and formal systems that assumes a clear distinction between syntax, i.e., rule-based operations on symbols which proceed independently of the symbols’ meaning, and semantics, i.e., operations based on the meanings of symbols. He maintains that computers are devices that exclusively employ syntactical operations. Given this point of view, Searle concludes that computers cannot think in the sense that a human being can; since computers cannot assign meanings, but only apply syntactical rules, they cannot possess understanding or mind. as we shall see below. Peirce, on the other hand, who was fascinated by the possibility of machine intelligence[1], explicitly repudiates such a distinction. Peirce's analysis of thought is radically different from Searle’s, in that he differentiates between three types of thought, all describable as variants of a basic syllogism; but maintains that only one is simple enough to be realized by computers. Although some machines, for Peirce, are extensions of people’s thought processes, and some are literally thinking machines, he concludes that the computer will never be able to think in a fully human manner, i.e., employing the full range of human capacities.

 

Peirce (Peirce, 1992c) offers detailed speculations and arguments (pp. 15-16) concerning the biology and physiology of thinking; these are based on some knowledge of neurophysiology and on speculation about the processes of the central nervous system (CNS). He argues that, starting with extremely complex combinations of very simple sensations, no more than the “excitation of a nerve” (p. 16), and continuing through various processes of inference, the mind creates conceptions of “mediate simplicity” (p. 16) which reduce the “inconceivably complicated” sets of sensations to simpler, abstract, concepts. These concepts are the bases and the components of our understanding of the world. To summarize one of his examples (p. 16), the motion of an image over the retina produces, through such inferential processes, the conception of space.

For Peirce, then, the elements of thought he termed “cognitions” were, roughly speaking, the operational building blocks of mind. Thus, in that same essay, he argues that cognitions, which he uses to refer to the contents (and, eventually, the objects) of thinking, are allied to sensation, and that there is no independent faculty of intuition. This latter faculty would involve the “determination” (p. 12) of a cognition “directly by the transcendental object” (p. 12). Peirce will argue that the interactions of these cognitions— our mental operations, constituting the basis for our inferences about the world and about our thinking — may be described in terms of one of several types of logical processes (Peirce, 1992d, pp. 28-55).

At this point, Peirce seems to have concluded that mind can result from formalizable operations on arbitrary symbols. That is, if a description of thinking may be couched in terms of logical processes (for further explication of these terms, see the next section), might these same processes be the equivalents of formal logical or algebraic procedures? If Peirce can relate thinking, in general, to cognitions whose interactions are describable in terms of formally manipulable inferential processes, the conclusion seems to be that this same cognition is based on a brain physiology describable by, perhaps even employing, formalizable operations. At first glance, then, his argument seems to start similarly to Searle’s Chinese Room argument (e.g., Searle, 1990), but to come to the opposite conclusion.

However, as will become clear, the above analysis is too superficial. Although Peirce’s position on the logic of thought and language differs radically from Searle’s, he arrives, in fact, at very similar conclusions.

 

The Chinese Room argument

 

A digital computer is a device that carries out processes by changing the values and spatial distributions of some physical entity, such as the voltages in various elements of semiconductor circuitry, in discontinuous steps in accordance with a finite set of well-defined (but not necessarily consistent) rules (in the case of modern computers, those rules are related to  - but not identical with - Boolean algebra, a logical calculus devised by the mathematician George Boole [Boole, 1958]). When numbers are represented in these devices, they are done so by a placeholder (digital) symbolism rather than by a quantitative (analog; see note 4) symbolism.

Searle, in his Chinese Room argument (Searle, 1994, pp. 200-220), argues that while the symbols' manipulations must be driven and instantiated by physical substances and processes, those physical processes have no relationship except the most abstract to the symbolic processes that they realize. His argument is twofold: first, that in these systems, since they employ abstract logic, the meanings of the symbols are entirely arbitrary: syntax is independent of semantics, and these systems’ rules are exclusively syntactical. Second, in the instantiation of this logic through the above physical processes, the relationship between any given physical quantity and the symbolic element it represents is also entirely arbitrary. In other words, the physical entities — the voltages, in this example — which comprise the functional elements of the device are, qua physical entities, irrelevant to the computer's function as a symbolic manipulator. In fact, it is quite possible, as Searle points out, to construct computers out of "cats and mice and cheese or levers or water…" (Searle, 1994, p. 207). As long as the dynamics of their relationships are constrained to correspond, at some level, to the syntactic relationships of symbolic logic, the actual physical realizations are irrelevant. In digital computers, he claims, all that is happening is the creation and alteration of strings of symbols that must be subsequently interpreted by a "minded" human being.

The Chinese Room argument, then, proceeds as follows: a person totally ignorant of Chinese sits in a closed room, and receives, through a slot in the wall, cards with Chinese characters on them. The person goes to a rulebook and finds some rule (or set of rules) relevant to the character (or the last few characters) just presented, and as a result of those rule-determined operations, picks some other Chinese character(s) from a pile, and passes it out through the wall. According to what Searle calls the "strong AI" position, "thinking is merely the manipulation of formal symbols" (Searle, 1990, p. 26). Thus, if the rules are complete, according to that position, a Chinese speaker will be able to hold intelligent conversation with the “room”: the totality of the operator, rules, and symbols. But Searle argues that there is no thing or person in the room that understands Chinese (nor does the room as a whole). Therefore, Searle concludes, even if that room could intelligently converse in Chinese it does so mindlessly, with no possible basis for understanding the symbols. Since, Searle argues, computers operate in this same fashion, solely on the basis of syntactic operations, they too are and must always be mindless. Thus, since "a physical state of a system is a computational state only relative to the assignment to that state of some computational… interpretation" (p. 210), one cannot generate mind from these arbitrarily instantiated formal processes. Searle notes that an observer who did not recognize Chinese characters, looking through a one-way window into the room, might understand the symbolic manipulations as stock-market formulas, and apply them consistently according to that interpretation (Searle, 1990, p. 31).

 

Peirce, logic, and thought

 

Several important issues raised by the above argument concern the nature of formalizability, of manipulations of symbols, and of the various types of formal logic. Peirce wrote voluminously on these subjects. Roughly speaking, according to Peirce (e.g., Peirce, 1992a), there are three basic types of logic, derived from the three-part syllogism. This syllogism consists of

 

R, a rule: (the beans in this bag are white),

C, a case of the rule: (these beans are from the bag),

E, a result: (these beans are white)

(Peirce, 1992a, p. 188).

 

By altering the order of the elements in this expression, Peirce realized that one could symbolize entirely different types of thinking. Thus, deduction consists of statements in the above order: (1) R, C, E; induction in the order (2) C, E, R; and hypothesis construction (also termed "abduction" (e.g., Houser & Kloesel, 1991, p. xxxviii; also Peirce, 1998b, p. 95), the order (3) R, E, C (Peirce, 1992a, pp. 188-189).

Before proceeding further, it is necessary to elaborate on Peirce's classification of types of thinking. "Firstness" has to do with "immediate feeling" (Peirce, 1992b, p. 260), a thinking involving only the "fleeting instant," which once past, is "totally and absolutely gone" (p. 259). These instants run in a “continuous stream through our lives” (p. 42). "Secondness" is type of thought in which the will appears, consisting "of something more than can be contained in an instant" (p. 260); the continuous stream of instants of thought begin to be combined by an “effective force behind consciousness” (p. 42). "Thirdness… is the consciousness of process, and this in the form of the sense of learning, of acquiring… the consciousness of synthesis" (p. 260).

Peirce then goes on to speak of three different senses of thirdness. The first is "accidental," and corresponds to "association by contiguity." This is interesting in its relation to behaviorism, to deductive logic, and also to our perception of space, for Peirce states that "we cannot choose how we will arrange our ideas in reference to time and space, but are compelled… [by an] exterior compulsion" (p. 261). The second type of thirdness is "where we think different things to be alike or different" (p. 261); a thinking in which "we are internally compelled to synthetise them or sunder them… association by resemblance" (p. 261). One is reminded of associational psychology, some aspects of cognitive psychology, and of inductive logic. The third type of thirdness is the highest, which the mind makes "in the interest of intelligibility… by introducing an idea not contained in the data" (p. 261). Here we have the kind of thinking involved with hypothesis construction and testing, with science in general, and with art. Peirce states, "The great difference between induction and hypothesis is, that the former infers the existence of phenomena such as we have observed in cases which are similar, while hypothesis supposes something of a different kind from what we have directly observed" (p. 197).[2] Peirce further states, "the work of the poet or novelist is not so utterly different from that of the scientific man" (p. 261). Habermas, commenting on abduction, states that, "what Peirce called 'percepts'… depend upon those limiting cases of abductive inference which strike us in the form of lightning insights" (Habermas, 1995, p. 254). Peirce held that the most important and creative aspect of the mind, i.e., "synthetic consciousness, or the sense of the process of learning" (Peirce, 1992b, p. 264), related to "thirdness" (p. 260) operates through a combination of habit and chance. It "has its physiological basis… in the power of taking habits" (p. 264). In addition, however, he maintains that "it is essential that there should be an element of chance… and that this chance or uncertainty shall not be entirely obliterated by the principle of habit" (p. 265). For Peirce there seems to be a kind of balance between habit: fixed thinking and behavior, and chance or uncertainty: fluctuations in thought and behavior, producing variants in fixed habitual patterns.

Peirce's stance on formal deductive logic is thus clearly delineated. He states that "formal logic centers its whole attention on the least important part of reasoning, a part so mechanical that it may be performed by a machine" (Ketner, 1988, p. 44). And we can now appreciate that this results from either or both of “firstness”, and thirdness of the first type, where nothing external to the given data is required or arrived at, and the sequence of association is "externally compelled." To put this explicitly in terms of the computer, this external compulsion is the equivalent of an algorithmic procedure, in which every step is determined in advance, i.e., every process realized by the physical machine is “compelled” through its programming.

We thus can relate the three types of syllogism above to the different kinds of thirdness. Deduction is related to the lowest kind, the association by contiguity, since it can proceed solely through habit, without choice, and can, in addition, be performed by machines. Induction, since it involves ideas of similarity and difference, and association by resemblance, would seem to employ the second type of thirdness; whereas abduction must employ the last type. This is a much richer classification of logic and thought than Searle's, insofar as the Chinese Room argument is concerned, and it opens several questions about the relationship between the two positions.

 

But Searle's question remains: despite its algorithmic nature, is this kind of thought "minded"? Does it require, or may it possess, understanding? Searle regards syntax and semantics as inherently separable in formal languages, and thus the difference between purely formal operations on symbols with arbitrary meaning is clearly differentiated from operations on symbols possessing meaning, i.e., semantic operations. In addition, while understanding may involve neural processes which are entirely unconscious but that may cause "conscious phenomena" (Searle, 1994, pp. 239-241), those neural processes are not mental processes. Unconscious neural processes can however correspond to, i.e., produce the same effects, as syntactic rules. Thus for Searle, only unconscious neural processes may be algorithmic; consciousness, dependent on but not identical with those neural processes, intrinsically involves meaning, and thus transcends algorithmic processes.

Peirce, on the other hand, regards all sensation as conscious. Thus, the entities he calls “sensations” at the lowest level seem to be mental entities that, although we may be aware of them in the instant in which they are present to us, we are not aware of as individuals over time. In fact, he claims (Peirce, 1992d, pp. 42-44) that those sensations are claimed to be “unanalyzable,” and “something which we cannot reflectively know” (Peirce, 1992d, p. 42).[3] However, as this quote indicates, although these sensations are immediate, fleeting, and gone "absolutely" when their instant of perception has passed, during the instant that they are present, they are mental contents. There is, in addition, “a real effective force behind consciousness” (my Italics; p. 42) — and note that this “force,” which works on the sensations, while it seems unconscious, is still, for Peirce, mental — which unites the individual sensations in their “continuous stream” into those mental events and contents of which we are continuously aware. Peirce, then, seems to have as his starting point a class of entities which correspond, roughly, to nerve signals which have something like the status of pre-qualia mental contents, a type of entity that Searle denies, if I understand him correctly.

Peirce's position on symbols derives from his analysis of mental contents. In contrast to Searle, he recognizes no symbols without meaning, i.e., related solely by syntactical rules. His extensive classification of types of signs — there are up to sixty-six types (Houser & Kloesel, 1991, p. xxxvii) — includes pictorial or analog symbols: icons which correspond structurally in some relevant manner to that which they symbolize.[4] Those icons, according to Peirce, are necessary for all thinking, including deductive logic. In addition, according to Ketner, Peirce uses the term "diagram" much more generally than to refer to pictures or pictograms, and instead refers to virtually all models — verbal, pictorial, or mathematical — as diagrams: "Indeed, 'diagram' for Peirce is roughly equivalent to a generalized notion of 'model'." (Ketner, 1988, p. 47). Thus, in speaking of errors which are made about the nature of logic, he states, "one such error is [in thinking] that demonstrative reasoning is something altogether unlike observation… [iconic representations] represent relations in the fact by analogous relation in the representation… it is by observation of diagrams that the [deductive] reasoning proceeds" (Ketner, 1988, p. 41).

The above serves to further explicate the differences between Peirce and Searle. Peirce was explicitly stating that deductive reasoning, as performed by mathematicians, involves much more than syntactical operations, in fact, that the representations necessary for deductive reasoning are, in part at least, not, as icons, arbitrary symbols. Even though we have seen that for him deductive reasoning is in part algorithmic, i.e., capable to some extent of being performed by machines, it necessitates, in toto, the highest levels of thought (e.g., "every kind of consciousness enters into cognition" [Peirce, 1992b, p. 260]). It is, then, not as a consequence of a fundamental difference in types of symbols and/or symbolic operations, with or without semantic content, but as a consequence of different levels of thought that Peirce argues for the non-computability of non-algorithmic logic.

Thus Peirce, in contrast to Searle, would not, I believe, allow any separation between syntax and semantics, in the following respect. He would claim that what  Searle is terming "syntactic rules" partake of what Searle would consider semantic characteristics, and, generally, that such rules must so partake. However, if those rules were simple enough so that pure deduction, i.e., thinking of the first type of thirdness, was all that was required, then a machine could indeed duplicate such "routine operations" (Peirce, 1992d, p. 43). In this simple sense, for Peirce, a digital computer has "mind" or "understanding." However, for Peirce, the crucial difference between man and machine lies in the use of "fixed methods" of thought, in contrast to "self-critical… formations of representations" (p. 46). "If a machine works according to a fixed principle… unless it is so contrived that… it would improve itself… it would not be, strictly speaking a reasoning-machine" (p. 46). The difference for Peirce, then, was not in a syntax/semantics difference, but in a difference in "self-control": "routine operations call for no self-control" (p. 43; also see below). That is, routine operations, which we now might term algorithmic procedures, while they may entail self-critical modifications and learning from error, nonetheless perform no hypothesis testing in the sense explicated above (the third sense of thirdness). In fact, Peirce states, referring to the difference between mental modeling in reasoning and laboratory experimentation with apparatus, that "every machine is a reasoning machine, in so much as there are certain relations between its parts… that were not expressly intended. A piece of apparatus for performing a physical or chemical experiment is also a reasoning machine… [they are] instruments of thought, or logical machines" (p. 52). A position similar to this, perhaps, is held today by Clark (Clark, 1998).

 

Difficulties

 

Each of these positions has its own problems. Searle's is very compelling, but it is based on a notion of symbol manipulation which is foreign to Peirce's conception, and in addition, a notion of understanding or meaningfulness which, while seeming plausible, even obvious, becomes upon investigation, if not problematic, at least extremely complex. That is, in the Chinese Room argument, Searle maintains that "like a computer, I manipulate symbols, but I attach no meaning to the symbols" (Searle, 1990, p. 26), and that in a computer, "symbols are manipulated without reference to any meanings" (p. 27). To maintain this, as I have mentioned, presupposes the clear separation of rule-governed operations, i.e., syntax, and meaning-governed operations, i.e., semantics. This partitioning has been disputed by theories in cognitive linguistics in which all linguistic operations encompass both semantic and syntactic components (e.g., Johnson, 1993; Lakoff, 1990). Natural language operations, in these theories, cannot be categorized in this manner, and neither, by extension, can operations in formal languages (see especially Lakoff & Núñez, 2000). Thus, for these theories, as for Peirce’s, even the simplest symbolic manipulation is, in some primitive sense, a manifestation of understanding.

In addition, one must note that the interpretation of the Chinese room “rulebook” is not actually arbitrary. That is, although Searle argues that the Chinese room might be interpreted as modeling the stock market (Searle, 1990), we must ask whether this argument actually holds. It sounds plausible, until one reflects that there must be some structural similarities between a symbolism and what it symbolizes. That is, all possible structures cannot be symbolized by the Chinese room’s syntactical operations. Thus, by inspecting sets of truth-tables corresponding to those operations, it is clear that some physical events simply will not fit their conditions, even though many, perhaps even an infinite number, will. One might then ask what subset of all possible structures are symbolized by some complex set of syntactical operations, and more importantly, what the implications of that limitation are. Once the door is opened to some limitation as to what a particular structure can symbolize, it seems to me that this particular argument of Searle’s has been, if not refuted, at least weakened: the interpretation of the symbolic operations in the Chinese Room is not arbitrary.

In yet another facet of this argument, Searle maintains that the “biological nature” (Searle, 1990) of the brain enables it to realize or generate meaning. However, we are (I will assume) material information-processors ourselves. As far as is known, if we view the central nervous system (CNS) microscopically we find only physical, mindless components: neural structures, various chemicals, and so forth, like the gears in Leibnitz’s mill. What assigns meaning to our internal operations (for more on this, see also Harnad, 1990 and the “symbolic anchoring” problem; Pattee, 1997, Cariani, 1989, and Cariani, 2001)? Searle argues that, unlike a computer, a biological system such as a real stomach digests real food. But this analogy does not carry through for the mind. We do not “digest” ideas, as entities, as our stomachs process food; we do not pump ideas, as entities, as our heart pumps blood. That is, ideas are simply not entities “floating around” to be snatched by mental forces, and then, by mental “chemicals”, somehow processed. We might use these expressions metaphorically, but not literally. But in that case, ideas or concepts are just as much abstractions for us as they are for the computer. A real robot, controlled by a computer, picks up a real block. If one maintains that the remote programmer is the ultimate interpreter in that situation, then one would have to say the same for the Chinese room, and that argument could not get started.

Peirce’s response, I believe, would be similar, and would include the point that the analog components of deductive symbolisms, the “diagrams,” do instantiate or imply a structuring analogous to the neuron’s physical structure (see note 4).

 

Peirce's ideas, however, also have their problems. There are systems with internal controls which cannot by any reasonable criteria be “minded”: thermostats, to take an extreme example, are such self-correcting machines. Their behavior, although it involves negative feedback — self-correction, in a sense — is, nonetheless, “externally compelled”: completely repetitive and determined, in the sense that it can never originate new states and merely repeats old ones. If a simple example like the thermostat clearly refutes Peirce’s general point, then it would seem that some other criteria, in addition to self-correction, are necessary. Precisely what, therefore, must "self-criticality" entail in a machine for it to think in any interesting way? Feedback and recursive processes in computers have been implemented for decades; computers now learn, in some sense of the term. Peirce might answer that computer learning proceeds through algorithms; that is, that the determinants of the changes in the computer's programming are themselves fixed, and thus at this level the machine is not self-critical, i.e., it cannot originate its own programming. The AI argument, in turn, might be that biological systems themselves are constrained by physiological parameters. Peirce's response, I believe, would be to invoke his principle of chance: "it is essential that there should be an element of chance in some sense as to how the cell shall discharge itself" (Peirce, 1992b, p. 265). A proponent of AI, however, might respond that in fact computers can and do use random number generators, for example, to produce just such an effect.

However, Peirce’s later writings (e.g., Peirce, 1998c, pp. 245-248)  seem to indicate that the relationship between what he termed “self-control” and the behavior of particular cellular structures is tentative, at best. The term “self-control” in this and similar essays refers to one’s willed actions and decisions. This is indeed a common meaning of the term, but its reference to the earlier analysis, in the context of biological structures, is not explicated. Given such an abstract meaning, it is fairly straightforward to claim that computers cannot act through the will, unless that faculty is equated, as I mention above, with mere random choice. But Peirce (especially the Peirce of this essay) would deny that latter option. Peirce stated, in fact, that in order to create a machine that can reason as a human can, it must “be endowed with a genuine power of self-control” (Peirce, 1998a, p. 387), which seems virtually impossible[5] to him. 

This latter position, however, because of the lack of an explicit relationship between biological structures and the will (hardly surprising, given the state of physiology at the time), leaves Peirce open, I believe, to claims by AI proponents that they have in fact caused machines to manifest willed action. A Peircian objection would be countered with an accusation of vagueness.

 

Peirce and Searle: conclusions

 

What are the implications of these differences between Peirce and Searle? We have seen that for all the initial superficial similarities, there are profound disparities between their attitudes toward symbolic systems and "thinking" machines. While Searle dismisses the possibility of computers which understand in anything like a human sense (unless those machines are "biologically based"), Peirce does not. For Peirce, virtually any machine is, first, an extension of human thought, and second, in the case of calculating machines, already employs some aspects of human thought and thus of human understanding. However, a Peircean “theorematic machine” (Ketner, 1988, p. 52) would require a kind of processing which Searle certainly, and Peirce probably, would regard as beyond the capabilities of the digital computer.

The latter point brings us back to the beginning of this essay, in which I indicated that I would introduce another argument with just that conclusion. Although moving through radically different terrain, both Peirce and Searle would, I think, agree with it.

That argument is the following:

 

An argument for the impossibility of mind in any finite-state device (or device in which the states are functionally partitioned into discrete or bounded subsets):

A) Let us assume that an automaton with only one possible behavioral path, i.e., with only one linear set of states, which could be infinitely long, cannot have mind.

B) Let us now add one contingency state: suppose that instead of the one path in A, there are now two paths, activated by two buttons or sensors, such that this device now instantiates a single counterfactual. That is, IF a red light shines on one sensor the device initiates one linear sequence OR IF a green light shines it initiates another. Now clearly either of these can be infinitely long; in addition, either could start at any point after the other has started. It is evident that this second device cannot have mind any more than the first; two mindless devices have merely been added together.

1. If this second sequence is initiated by some internal state change so that the contingency now depends on an internal switch instead of an external one: the number of times, for example, that the device passes through the first state sequence, we have still merely added two mindless devices together.

C) Through induction we can now see that any finite-state device cannot have mind. Further, it would similarly seem that any device with a countably infinite number of states (initial states, i.e., a countably infinite-state device in the sense that a finite-state device is finite) cannot have mind. [6]

 

One could object to the above argument, however, on the following grounds: that it is valid only if the state-paths do not change dependent on context. That is, when a contingency is added, if the original linear set of states remains the same, clearly we are dealing with a simple additive linear process. However, if the original state sequence changes when a second is added for another contingency, then the additive process is no longer linear. A computer programmed along those lines might refute this argument, if the change took place in both the original and the new sequences. In that case, the interdependence of the sequences might invalidate the linearity criterion.

To put this another way, this counter-argument might be termed a “lemon-meringue” example, i.e., a synergistic or retroductive counter-argument. If the system acted, in effect, as a gestalt, where any change to any part changed all (or virtually all) other parts, then linearity no longer applies. Mind then would be a synergistic property, an interactive property, similar to the emergence of a lemon-meringue pie from lemons, eggs, etc., where the resulting whole has none (or few) of the properties of the original components. That is, if this non-linear case held, an inductive argument would not be appropriate to critique it, since “mind” (or better, perhaps, “the mental”) does not arise - since it is not a property (as in the blackness of all crows) of individuals - from the sharing of a property over a class of individuals, as in classical induction. Thus, the inductive argument would be misapplied to refute a retroductive process (and this is perhaps an interesting comment on the limitations of induction).

The lesson here, however, is that if the above argument is correct, a computer program which does not work synergistically or retroductively, i.e., which does not change its components with context, as does the CNS, will not generate mind.

 

At this point, there are some indications as to the precise differences between biological systems and digital computers which might enable the latter to embody meaning, but those differences do not hinge on intrinsically biological characteristics; they can be realized by machines.

Let us look more closely at the mechanics of the CNS. I will argue that the well-known analogy between the CNS and the digital computer is, in fact, almost totally incorrect. Roughly speaking, the CNS operates as follows: sensory organs excite neurons such that the frequency of the impulses resulting from stimulation is proportional to the (logarithm of the) magnitude or intensity of the stimulation. Thus an analog property, frequency, encodes magnitude from the most peripheral aspect of the CNS. When these impulses, carried along axons, enter processing areas (the visual cortex, let us say) of the CNS, many inputs converge on single cells. So far, some “processing” has occurred (starting from the initial transduction of light to neural impulses, frequency-coded, through a complex series of procedures on the retina), leading to the transmission of impulses further into the CNS. When a subset of these early impulses arrive at a presynaptic terminal of a cell in a cortical area, they are again transduced into chemical transmitters (except for some few neurons connected directly through electrical fields or ions) which rapidly diffuse across the synaptic gap. These chemicals then depolarize (unless they are inhibitory, then they hyperpolarize) the post-synaptic membrane. This depolarization then spreads rapidly over the surface of the membrane, reaches a maximum in magnitude (an analog property) and area (another analog property) in some short time period (another analog property), then recedes. During this period, if the magnitude of the depolarization is great enough, this neuron will fire (employing, as before, frequency coding). If it is not great enough, it will not. Meanwhile, other areas of the membrane (and there may be thousands) on the cell body of the same neuron are being depolarized similarly (and their terminals may themselves be moving around on the cell body, disappearing or appearing). If the first area of depolarization, while it is expanding or contracting, is intersected by other areas, caused by transmitters from other terminals, and the sum (or product, or some other function – no one actually knows) of those depolarizations exceeds some (variable) threshold, then that neuron will fire. It is this latter process, consisting of the dynamic interaction of many areas of de- and hyper- polarization on the surface of the cell body (which may extend into a huge dendritic tree), which constitutes, for the most part, neural processing (as it is understood today – although I have not even mentioned such aspects as different neurotransmitters, effects of memory, and many others). There is no digital coding – i.e., a placeholder code employing a finite set of arbitrary basic elements - whatsoever in this huge set of processes (and the above concerns only one neuron – I have not dealt with “cell-assemblies”); it is all accomplished through what may indeed be compared to Peircean “diagrams”: analog “representations” (and I employ this term very hesitantly), in which some aspect of the real-world structure of the representation corresponds to that which it represents.

A Peircean "theorematic" machine, then, might be constructed from a model similar to the above. We might expect that it would employ “diagrammatic” symbols instead of digital symbols. These would possess structure, as physical entities, which corresponded to what they symbolized: they would be analog symbols. In addition, just as the structure corresponds, the manipulations of those symbols would have to correspond, to some extent, with the manipulations of what was symbolized. Taking the example of the slide rule, not only does the structure of the “slide” correspond to that of the number line, but the operation of sliding one piece past another literally increases or decreases the length of the rule, as the quantities it represents increase or decrease. Even such a simple analog device does not employ a finite set of well-defined symbols, and as far as its operations are concerned, in the case of a slide rule, they are finite and well-defined. However, we might consider the more general case in which the symbols (or pre-symbolic entities, e.g., the spatial patterns of depolarization on the neural cell body surface) employed possess complex analog structures, and the operations in which they engage are based on those structural characteristics. In general, even if there were clear rules having to do, say, with how differently-shaped constituents of these symbols interacted, the specifics of those shapes could vary enough to alter that interaction enough so that the rule would have to be considered to be changed. That is, there might be rules that held on a microstructural level determining interactions of the “atoms” of some analog entity or structure; yet because of variability and complexity in the way in which microstructures were combined, the regularities from which rules of the “macrostructure” of the analog entity were inferred might only be approximate. To put it more generally, if “rules” were regularities corresponding to attractor basins in a “rule state-space” generated through the dynamics of complex analog interactions, then to speak of “well-defined” rules would be to approximate the actual state of affairs. Yet this picture is not unreasonable in the CNS, given the variability and complexity of the central nervous system.

It seems possible, then, to construct devices for which assumption B of the argument above, where it involves the linearity and independence of the combining processes, does not hold (one might also speculate, here, about extensions to Turing machine theory). These devices, in addition, seem consistent with Peirce’s requirements for a lack of “external compulsion” on a “theorematic machine”, since their operations would in part determined by momentary configurations of their components, set by external input, internal noise, and/or through states too complex to be computable. Further, since these devices would at least in part consist of analog components, the “fine-grainedness” of their material parts would be relevant to their construction, and Searle’s requirement (for a system similar – in this respect – to a biological entity) would also be satisfied.

Throughout the literature, we have see paeans of praise for the digital computer as the eventual embodiment of mind (e.g., Dennett, 1991). We have also seen severe critiques of this position (e.g., Dreyfus, 1993). In the former case, these arguments against the critics have been almost irrelevant, in the sense that efforts in the field of AI have continued unabated. In the latter case, very few constructive suggestions as to how to proceed, given that the arguments against conventional AI were correct, have been offered. I have described one very abstractly above; I will put forward a concrete example below.

Have there been devices similar to the above analog systems constructed? A few examples come to mind; the first steps, so to speak, in this direction: the very early description of an analog “homeostat” by Ashby (Ashby, 1960); the modern example of the “silicon retina” (Mead, 1988; Harrison, 2000), which has been developed to the point where it is being employed medically. But these are not computers in the general sense that they manipulate abstract symbolic operations. The problem, I believe, is the following: the current notion of a “computer” relies on the implicit assumption that it is the operations themselves that are important; that logic, however it is formulated, even as fuzzy logic, must be manipulated in order to reach goals. Even “evolutionary” computer systems, creating and discarding heuristics, are centered on just that class of operation. And given the construction of the digital computer, in which the must fundamental physical components instantiate truth-tables (Boolean or not), what other choice is there?

However, we might consider a system like the CNS, not as an “approximation” to a digital computer, but as an example of a system which turns this viewpoint around. That is, instead of keeping constant the fundamental logical processes and combining them in different ways to achieve goals, would it not be possible to employ the goals themselves to alter the logical parameters, to change, not merely the truth-tables, but the operations comprising those tables; the physical parameters of the system? This is not possible in a digital computer (except perhaps in high-order “virtual” machines – and the debate as to whether such approximations serve as well as actual devices is ongoing), where the circuitry is composed of transistors that embody binary operations; and AND, OR, and NOT, or similar, gates are structures imposed on the silicon matrix.

Suppose, however, that a machine were built in which those fundamental physical structures were themselves variable, as are the neural structures in the CNS. That is, suppose that the physical components which instantiate the truth-tables, now realized as transistors whose outputs have binary read-out imposed on the actual analog signal, memory components read similarly, and the AND, OR, etc., operations, realized as circuitry operating according to similar rigid filtering, were instead comprised of neural nets or the equivalent. Then the operations which were previously fundamental would be realized at a finer level of structure, in effect, and become variables, reorienting themselves as necessary to realize the system’s goals. More specifically, suppose that one such neural net were initially trained to duplicate the responses of a logic circuit realizing a truth-table of some sort. Although the initial responses of that net would duplicate those of that table, it might a) routinely vary somewhat, and could b) radically alter its operation if the necessities of the overall system required. Given that the components of that net were analog, as in the VSLI structures of the silicon retina or the components of the CNS, we would have a device which might be described as “picturing” the logic circuit, analogous to the way that a diagram pictures its referent. And that picture could change, suddenly or gradually, depending on the requirements of the task, given that the totality of neural nets realizing the logical operations were interconnected such that a synergistic or non-additive system, as generally described above, was created. This type of general-purpose analog computer realizing, in effect, a virtual digital computer, is probably just at the edge of present technology, but certainly not theoretically impossible.

 

In this essay I have attempted both criticism and an explicit indication of a direction for AI research. The positions critical of digitally-implemented AI do have, I believe, very strong arguments in their favor; yet in order to make an impact on the general endeavor of creating “minded” devices, an endeavor which should be ultimately successful if a materialist position on mind is at all reasonable, positive and constructive efforts to direct that effort need to be made. In this essay the suggestion for such a direction is made in the realization that in the ongoing creation of analog/digital robot/computer hybrids, in analog sensory networks, and in the advancement of analog processing, the criticisms offered to AI are, I believe, beginning to be answered.

 


References

Ashby, W. R. (1960). Design for a brain: the origin of adaptive behaviour. London: Chapman and Hall Ltd.

Boole, G. (1958). An investigation of the laws of thought (2nd ed.). New York, NY: Dover Publications.

Cariani, P. (1989). On the design of devices with emergent semantic functions. State University of New York, Binghamton, NY.

Cariani, P. (2001). Symbols and dynamics in the brain. BioSystems, 60, 59–83.

Clark, A. (1998). Being there: putting brain, body, and world together again. Cambridge, MA: The MIT Press.

da Silveira, L. F. B. (1999). Some considerations about semiotic machines from the point of view of Charles S. Peirce's Philosophy, [http://www.inm.de/kip/SEMIOTIC/silveira_article.html]. ARISBE: http://www.door.net/arisbeJune].

Dennett, D. C. (1991). Consciousness explained (1st ed.). Boston, MA: Little, Brown and Company.

Dreyfus, H. L. (1972). What computers can't do. Cambridge, MA: The MIT Press.

Dreyfus, H. L. (1993). What computers still can't do: a critique of artificial reason (2nd ed.). Cambridge, MA: The MIT Press.

Habermas, J. (1995). Peirce and communication. In K. L. Ketner (Ed.) (1st ed., Vol. 1). New York, NY: Fordham University Press.

Hanson, N. R. (1965). Patterns of discovery; an inquiry into the conceptual foundations of science. Cambridge [Eng.]: Cambridge University Press.

Harnad, S. (1990). The symbol grounding problem. Physica D, 42, 335-346.

Harrison, R. R. a. K., C. (2000). A silicon implementation of the fly's optomotor control system. Neural Computation, 12, 2291-2304.

Houser, N., & Kloesel, C. (1991). The essential Peirce: selected philosophical writings (Vol. 1). Bloomington, IN: Indiana University Press.

Johnson, M. (1993). Moral imagination: implications of cognitive science for ethics (1st ed.). Chicago: The University of Chicago Press.

Ketner, K. L. (1988). Peirce and Turing: comparisons and conjectures. Semiotica, 68(1/2), 33-61.

Lakoff, G. (1990). Women, fire, and dangerous things (2nd ed.). Chicago, IL: The University of Chicago Press.

Lakoff, G., & Núñez, R. E. (2000). Where mathematics comes from: how the embodied mind brings mathematics into being. New York, NY: Basic Books.

Mead, C. (1988). Analog VLSI and neural systems. New York, NY: Addison Wesley Longman, Inc.

Pattee, H. H. (1997). The physics of symbols and the evolution of semiotic control, Workshop on control mechanisms for complex systems: issues of measurement and semiotic analysis, Las Cruces, New Mexico, Dec. 8-12,1996 (Vol. Proceedings Volume). Redwood City, CA: Addison-Wesley.

Peirce, C. S. (1992a). Deduction, induction, and hypothesis. In N. Houser & C. Kloesel (Eds.), The essential Peirce: selected philosophical writings (Vol. I, pp. 186-199). Bloomington, IN: Indiana University Press.

Peirce, C. S. (1992b). A guess at the riddle. In N. Houser & C. Kloesel (Eds.), The essential Peirce: selected philosophical writings (Vol. I, pp. 245-279). Bloomington, IN: Indiana University Press.

Peirce, C. S. (1992c). Questions concerning certain faculties claimed for man. In N. Houser & C. Kloesel (Eds.), The essential Peirce: selected philosophical writings (Vol. I, pp. 11-27). Bloomington, IN: Indiana University Press.

Peirce, C. S. (1992d). Some consequences of four incapacities. In N. Houser & C. Kloesel (Eds.), The essential Peirce: selected philosophical writings (Vol. I, pp. 28-55). Bloomington, IN: Indiana University Press.

Peirce, C. S. (1998a). The basis of pragmaticism in the normative sciences. In N. Houser & A. De Tienne & J. R. Eller & C. L. Clark & A. C. Lewis & D. B. Davis (Eds.), The essential Peirce: selected philosophical writings (Vol. II, pp. 360-371). Bloomington, IN: Indiana University Press.

Peirce, C. S. (1998b). On the logic of drawing history from ancient documents, especially from testimonies. In N. Houser & A. De Tienne & J. R. Eller & C. L. Clark & A. C. Lewis & D. B. Davis (Eds.), The essential Peirce: selected philosophical writings (Vol. II, pp. 75-114). Bloomington, IN: Indiana University Press.

Peirce, C. S. (1998c). What makes a reasoning sound? In N. Houser & A. De Tienne & J. R. Eller & C. L. Clark & A. C. Lewis & D. B. Davis (Eds.), The essential Peirce: selected philosophical writings (Vol. II, pp. 242-257). Bloomington, IN: Indiana University Press.

Searle, J. R. (1990). Is the brain's mind a computer program? Scientific American, 262(1), 26-37.

Searle, J. R. (1994). The rediscovery of the mind (5th ed.). Cambridge, MA: The MIT Press.

Wiener, P. P. (Ed.). (1951). Leibniz, selections. New York: Charles Scribner's Sons.



[1] "It is widely acknowledged the intense interest Peirce manifested on logical machines since his father tried to install a Babbage's one at the Albany Observatory. During the period he stayed at the Johns Hopkins University he had Allan Marquand as one of his best students, and when this latter proposed to construct a logical machine, Peirce contributed with all his knowledge in Logic and Physics to the best execution of the project. But while he was an enthusiast of the computational findings with machines, he was very critical about the possibilities of a machine showing genuine logical reasoning." (da Silveira, 1999).

[2] Hanson terms this type of thinking “retroduction” (Hanson, 1965).

[3] This position might be derived from his reading of Leibnitz: "noticeable perceptions also come by degrees from those which are too minute to be noticed" (Wiener, 1951, p. 378).

[4] That is, the crucial difference between analog and digital symbolisms is that analog symbols, as physical entities, share structure with what they stand for. For example, in drawing a triangle, the colors of the lines are usually irrelevant, but their number is always relevant; a triangle has three sides, and so must the drawing of a triangle. Similarly, the numbers on a slide rule stand for the physical distance on that rule, and that distance is utilized to realize slide rule computation.

[5] “…we have as little hopes of doing that as we have of endowing a machine made of inorganic materials with life” (Peirce, 1998a, p. 387).

[6] This is not a sorites argument. Such an argument brings traits of the conclusion into the premises or original statements. That is, the classic induction: one crow is black, two are black, therefore all are black says nothing in any of the individual statements about any general characteristics or conclusions. The sorites argument, in contrast, in which one says: the removal of one hair does not bring about baldness, of two hairs does not, therefore the removal of all hairs does not, brings into the individual statements the general characteristic that must be demonstrated, and thus involves a kind of circularity. However the argument about the linear sequence of states is inductive; the individual statements about the absence of mind in any linear sequence of states make no claims that involve the conclusion, i.e., that involve the collective set of states, as does a sorites argument.

In addition, to claim that this is a sorites argument because of vagueness begs the question. That is, that objection assumes that mind will result if enough processes are added together. But that conclusion is just what this argument is aimed against, and of course assuming it at the outset will make this inductive argument untenable.