WHOLES AND THEIR PARTS IN COGNITIVE PSYCHOLOGY: SYSTEMS, SUBSYSTEMS, AND PERSONS © Anthony P. Atkinson Psychology Group, King Alfred's University College, Winchester, Hampshire, SO22 4NR, U.K. email: atkinsona@wkac.ac.uk Paper presented at the conference "Wholes and their Parts", Maretsch Castle, Bolzano, Italy, 17-19 June, 1998. http://www.soc.unitn.it/dsrs/IMC/IMC.htm
ABSTRACT Decompositional analysis is the process of constructing explanations of the characteristics of whole systems in terms of characteristics of parts of those whole systems. Cognitive psychology is an endeavour that develops explanations of the capacities of the human organism in terms of descriptions of the brain's functionally defined information-processing components. This paper details the nature of this explanatory strategy, known as functional analysis. Functional analysis is contrasted with two other varieties of decompositional analysis, namely, structural analysis and capacity analysis. After an examination of these three varieties of analysis, there follows a consideration of a mistake to avoid when conducting decompositional analyses in psychology, and a possible limitation on their explanatory scope. ----------------------- WHOLES AND THEIR PARTS IN COGNITIVE PSYCHOLOGY: SYSTEMS, SUBSYSTEMS, AND PERSONS [1] My aim in this paper is to elucidate the central explanatory strategy of cognitive psychology. My account owes much to the seminal work of Cummins (1983), but whereas he wrote from the perspective of a philosopher of science, my elucidation of the strategy is from the perspective of a practising cognitive psychologist. My hope is to cast the strategy in a somewhat new light. [2] In Section 1, I briefly introduce the notion that much psychological explanation, especially cognitive psychological explanation, is concerned with the search for the bases of dispositions. In Sections 2 and 3, I elaborate this notion into an account of the central explanatory strategy of cognitive psychology, which I call decompositional analysis. In this account, I distinguish two explanatory paths, each of which can be summarized by a distinctive type of question one can ask of a whole system. Given that a system has a certain property or capacity, one can ask: (1) What other properties or capacities does the whole system have in virtue of which it has the first property or capacity?; and (2) What is the material and/or functional composition of the system in virtue of which it has the specified property or capacity? I call the first explanatory path capacity analysis, and the second componential analysis. I shall distinguish two specific types of componential analysis, one of which is concerned with specifying the component functions or processes of a system, the other of which is concerned with specifying the material components or substrate of a system. In Section 4, I close the paper by briefly considering a mistake to avoid when conducting decompositional analyses in psychology, and a possible limitation on their explanatory scope. 1. DISPOSITIONAL REGULARITIES AND THEIR BASES Statements of dispositional regularities are a common form of ceteris paribus (i.e., all else being equal) laws. The standard view of ceteris paribus laws is that they are laws that apply to certain classes of entity some of the time, that is, when and only when certain specified conditions are met. [3] For example, it is a causal law that salt dissolves in water, ceteris paribus (i.e., there are exceptions to the law that salt dissolves in water). Statements of dispositional regularities are generalizations of the form, system S does X or exhibits Y under such-and-such conditions. Or as Cummins (1983, p.18) puts it:
"To attribute a disposition d to an object x is to assert that the behavior of x is subject to (exhibits or would exhibit) a certain lawlike regularity: to say x has d is to say that x would manifest d ... were any of a certain range of events to occur."
For example, human skin has the disposition to turn red and peel after exposure to ultraviolet light, over a certain range of conditions (e.g., amount of melanin in the individual's skin, and exposure time). What is distinctive about dispositional regularities (and possibly ceteris paribus laws in general) is that one is entitled to ask, of the system in question, what it is in virtue of which that system has the specified disposition. Strict or exceptionless laws, in contrast, are brute; there is no further question as to how strict laws operate - they just do, because that is the way the world is. In-depth explanation of dispositional regularities, of how and why a given system has certain dispositions, requires a search for more intrinsic properties in virtue of which the system in question has those dispositions. In other words, dispositional regularities obtain in virtue of certain distinctive facts about - in particular, properties of - the type of system that has them. As Cummins puts the point (1983, p.18):
"The regularity associated with a disposition is a regularity that is special to the behavior of a certain kind of object and obtains in virtue of some special facts about that kind of object."
An answer to the question of why salt is water soluble, for example, will appeal to the properties of the molecular structure of salt. Salt manifests the disposition to dissolve in water in virtue of having a certain sort of molecular composition. Likewise with many other dispositional properties: an answer to the question of why a system manifests a disposition will typically consist in a specification of that system's component parts, their functions and organization; a specification of, in short, that system's microstructure. (For more on the nature of dispositions, and of dispositional explanations, see, for example, the collected essays in Crane, 1996.) The idea that adequate explanations of the dispositional regularities of complex entities require appeal to the components or microstructure of those entities is nicely captured by Fodor's (e.g., 1974, 1989) views concerning the contrast between the "basic sciences" (i.e., the various branches of physics) and the "special sciences" (e.g., geology, chemistry, psychology, economics). Both the basic and the special sciences are, on Fodor's view, in the business of providing laws that explain their proprietary phenomena. Basic science laws are basic in the sense that nothing more is needed for causal explanation than for it to be shown that the phenomena are subsumed under one or more fundamental laws. In the case of special science laws, in contrast, there will always be a further story to tell about how and why those special laws apply. It is just a brute fact of the nature of the universe that basic science laws operate, but it is not a brute fact that special science laws operate. So the buck of special science explanations does not stop with a statement of the appropriate laws. Special science laws operate only in virtue of the reliable or proper functioning of mechanisms that implement these laws, so explanations in the special sciences have to be backed up by theories about how the operations of these mechanisms implement these laws. Adequate explanations of behavioural and property dispositions are accounts of how, or what it is in virtue of which, systems have those dispositions. The explanatory strategy that generates accounts of the bases of dispositions is what I am calling decompositional analysis. 2. DECOMPOSITIONAL ANALYSIS AS A RESEARCH STRATEGY IN COGNITIVE PSYCHOLOGY Decompositional analysis is the analysis of systems in terms of their subsystems. Rather obviously, the notion of a system and the notion of a subsystem are coupled. A subsystem is a component of a system. A certain sort of circuit board is a subsystem of a CD player, and heart valves are subsystems of hearts. Moreover, many single system-subsystem pairs will themselves be a part of a chain of such pairs. That is to say, there will be a hierarchy of subsystem-system relations, such that a system can itself be a subsystem of a larger system. So a certain sort of circuit board is a subsystem of a CD player, which is in turn a subsystem of a hi-fi system. And heart valves are subsystems of hearts, which are in turn subsystems of circulatory systems, which are in turn subsystems of organisms. Likewise, neurons are subsystems of neural circuits, which are in turn subsystems of the brain. [4] This way of viewing a system as itself a subsystem of a larger system need not stop with medium-sized objects such as hi-fi systems and organisms. A galaxy is a system composed of stars and planets, and galaxies are in turn subsystems of the universe. Similarly, atoms are systems composed of subatomic particles, but atoms are also subsystems of molecules. Decompositional analysis is a powerful and widely used method in many sciences for explaining the properties and capacities of complex systems. Its central ideas underpinned the development of the mechanical and engineering sciences, and its use is especially evident in the biological and psychological sciences (Bechtel & Richardson, 1993; Wimsatt, 1976). Indeed, it was in the mechanical and engineering sciences that decompositional analysis was crystallized and fully fostered. Not only was the behaviour of machines, and later, electronic computers, more easily disclosed and understood from the perspective of this mode of explanation, but it was also the basic principle upon which these machines were designed (e.g., Wimsatt, 1976). So to the extent that this explanatory strategy has been central to cognitive psychology, the engineering and computing sciences have been significant historical contributors to that discipline (e.g., Knapp, 1986). Indeed, cognitive psychological explanation, qua decompositional analysis, is sometimes described as a case of 'reverse engineering' (e.g., Dennett, 1990, 1994, 1995). Engineers construct machines from a plan or design that specifies what the machine is to do and how it is to do it, including what components it will have such that it is able to do those things. A reverse engineer, in contrast, tries to work out the largely or entirely unknown design of a machine by analyzing or decomposing it. A reverse engineer supposes that the machine in question is composed of parts, and that those parts realize certain functions. She attempts to explain how it is that the system does what it does in terms of its parts and/or their respective functions. Via decompositional analysis, one can explain how it is that a system exhibits certain dispositional regularities, that is, explain what it is in virtue of which a system has certain capacities and properties. The central idea of this scheme is that characteristics of whole systems can be explained in terms of characteristics of parts of those whole systems. (Similarly, performances of whole tasks can be explained in terms of performances of parts of those tasks, and complex events can be understood as series of connected simpler events.) This explanatory task is described by Cummins (1983) as consisting of an analysis of a system, S, that explains what it is for that system to have a capacity or property, P, "by appeal to the properties of S's components and their mode of organization" (p.15). Cummins continues (1983, p.15):
"The process often has as a preliminary stage an analysis of P itself into properties of S or S's components. This step will loom large ... [in cases where the Ps are] complex dispositional properties such as information-processing capacities." My task in this section is to elucidate the way this explanatory methodology is employed in cognitive psychology. In what follows, I shall refer to the field of cognitive psychology as that branch of psychology that is characterized by the information-processing approach. This is a mechanistic approach to psychology whose primary object of study is the nervous system, and in particular, the brain, viewed as an abstract information-processing system. As such, the information-processing approach should be distinguished from another strand of cognitive psychology, which is the study of cognition and perception, strictly so called. What characterizes the second strand is the emphasis on psychological capacities and properties of persons, such as perception, thinking, reasoning and problem solving. Cognitive psychologists of this second stripe seek explanations of such properties and capacities of people in terms of other capacities and properties of those people. So, for example, Tversky and Kahneman (1974) investigated how people reason under uncertainty. They sought to explain how "people assess the probability of an uncertain event or the value of an uncertain quantity" (p.1124). Tversky and Kahneman suggested that
"people rely on a limited number of heuristic principles which reduce the complex tasks of assessing probabilities and predicting values to simpler judgmental operations" (p.1124).
One such proposed heuristic is that of "availability", according to which
"people assess the frequency of a class or the probability of an event by the ease with which instances or occurrences can be brought to mind" (p.1127).
(The distinction between the level at which persons are characterized and the level at which parts of persons, such as the brain and its components, are characterized, will be discussed in Section 4.) In contrast to proposals of this sort, information-processing psychologists seek to explain how it is that humans have certain capacities and properties and can engage in certain activities, including those capacities, properties and activities studied by cognitive psychologists of the second stripe (hence the two strands of cognitive psychology are somewhat interrelated). Instead of asking questions about what people do, such as what strategies are used when reasoning under uncertainty, information-processing psychologists ask questions such as: What faculties or mechanisms underlie the ability to reason? How is it that humans (and other animals) can see, hear, feel, taste and smell the world in the way that they do? How is it that humans so readily develop the ability to speak and understand, and usually read and write, language? How is it that people can recognize the faces of family and friends? The information-processing psychologist answers such 'How is it that ...?' questions by saying roughly this: "People have these capacities and properties in virtue of having a brain that is principally an information processor. The brain enables a person to have these capacities and properties in virtue of its having such-and-such information-processing components operating in such-and-such a way." Answers of this sort are proposals about the functional architecture of the mind or brain. While the brain is the central object of study of the information-processing approach, psychologists who take this approach tend not to look at the structures and operations of the brain directly. Rather, they proffer theories about the information-processing architecture of the brain that are somewhat abstracted from neuroanatomy and neurophysiology. The thought is that a legitimate and useful form of psychological theory is one that attempts to explain the capacities and behaviour of humans in terms of a view of the brain as a complex of subsystems that are defined by what they do, not by what they are made of.
Consider one well-known model of face processing from the mid-1980s (Bruce & Young, 1986). The authors depict this model as a 'box-and-arrow' diagram (Figure 1). Box- and-arrow diagrams or flowcharts are very useful devices for depicting the individual stages of a process, and the order or flow of these stages. The boxes depict operations, processes or functions, but the theorist typically makes few if any commitments as to what it is that performs those functions. When such diagrams are used to depict the operations of information-processing systems, the boxes represent the functions of information processors, memory stores, sensory transducers, output mechanisms, and the like. For example, the 'expression analysis' box and the 'facial speech analysis' box of Figure 1 are functionally defined information processors, the 'face recognition units' and 'person identity nodes' boxes are essentially functionally defined memory mechanisms, and the 'name generation' box is a functionally defined output mechanism. Accordingly, the arrows in this type of model represent the flow of information between information-processing components, that is, they specify which outputs get to be which inputs. In other words, arrows represent the order in which the identified processes or functions are or can be done: one process comes after another because it depends on the output of the other process. In short, cognitive psychology is information-processing psychology. On this view, the canonical conception of the internal structure and operation of the brain as an information processor is of a structured system of component information processors, memory stores, sensory transducers and the like. What I shall now do is elucidate my view that this explanatory strategy in cognitive psychology consists of the use of three types of decompositional analysis. [5] In doing so, I shall highlight the dominant role of one of these varieties, namely, functional analysis. 3. VARIETIES OF DECOMPOSITIONAL ANALYSIS There are two (ideally interacting) explanatory paths within decompositional analysis, as follows (see Figure 2).
&nbs p; (1) Capacity analysis. A capacity analysis involves the specification of a capacity of a whole system, and the parsing of this capacity into subcapacities of the whole system. A capacity analysis is an attempt to answer the question, How does a creature perform a complex task? The general form of the answer provided by a capacity analysis is that the whole creature performs the complex task by virtue of performing a set of simpler tasks. In short, a task that a whole system has the ability to perform is decomposed into subtasks that that whole system has the ability to perform. So, for instance, the task of baking a cake might get decomposed into a number of more specific tasks that the cake baker performs, such as measuring the ingredients and then mixing them. Another example: An organism's capacity to secure a mate might get broken down into the capacities of producing a mating call, showing off some plumage, and fending off rivals. How do you bake a cake? You read and follow a recipe which tells you to mix such- and-such ingredients in such-and-such quantities and then put the mixture in an oven at such- and-such a temperature. This is a story about you, about you as a reader and follower of a recipe, about you as a baker of cakes. The details of that story are not parasitic upon an account of the information-processing mechanisms in your brain that enable you to read the recipe and bake the cake. That is, although there must be some story to be told about the information-processing machinery that enables you to do these tasks, the system-level story does not necessitate any particular account of the information-processing substrate. Similarly for other examples of capacity analyses: in general, a capacity analysis is a descriptive and explanatory story about a system, without reference to that system's subsystems. That is, capacity analyses are neutral with respect to mechanism. As Cummins says (1983, pp.29-30):
"Since we do this sort of analysis [i.e., capacity analysis, in my terminology] without reference to an instantiating system, the analysis is evidently not an analysis of an instantiating system. The analyzing capacities are conceived as capacities of the whole system. ... My capacity to multiply 27 times 32 analyzes into the capacity to multiply 2 times 7, to add 5 and 1, etc. These capacities are not (as far as is known) capacities of my components; indeed, this analysis seems to put no constraints at all on my componential analysis." (2) Componential analysis. The starting point for a componential analysis is some capacity or set of capacities of a whole system, and the goal is to explain how it is that the whole system has that capacity or set of capacities in virtue of having certain component mechanisms with a particular structure and operation. So a componential analysis starts from a look at a whole system and moves to hypotheses about that system's parts. At the same time, causal explanation is seen to proceed in the opposite direction, from parts to wholes. That is to say, the analysis of a system into subsystems is grounded on the assumption that those subsystems act and interact to cause the whole system's behaviour. There are two types of componential analysis (see Figure 2). To attempt to delineate a complex of processes or functions performed by one or more as yet unspecified material components of a system is to engage in (2a) functional analysis. And to attempt to delineate the material components or substrate of a system is to engage in (2b) structural analysis. Although these two types of componential analysis are distinct, there is, in practice, a close interplay between them (as indicated by the arrows in Figure 2). Functional analysis is the type of decompositional analysis that produces the standard form of box-and-arrow diagrams in cognitive psychology. A functional analysis is a specification of (i) what gets done by one or more components of a system, without commitment as to the material composition of those components, and (ii) the order of and relations between these functions or effects (i.e., a process, where the specified functions or effects are stages in that process). As such, a functional analysis can serve as an explanation of how it is that the system under investigation has a certain capacity or property. More specifically, a functional analysis provides an explanation of a system-level capacity or property insofar as it proposes that the system of interest has a certain functional architecture that enables it to have the identified capacity or property. A structural analysis, in contrast, involves breaking down a system into subsystems defined by their material structure, and specifying the organization of and relations between these subsystems. Structural analyses yield descriptions of the material composition of a system, and it is to such descriptions that one must turn if one wants to know what material apparatus implements the functions and processes detailed in a functional analysis, and how it does so. (Bearing in mind, of course, the above point that a structural analysis might cut at very different boundaries from a functional analysis.) Neuroscientists are in the business of carrying out structural analyses of the brain. (Though most are in the business of carrying out functional analyses too; that is, structural and functional analyses tend to be intertwined in neuroscientific research.) For example, many of the structures that underlie vision have been identified, including separate cortical areas specialized for such visual capacities as the perception of depth, form, colour and movement (see e.g., Cowey, 1985; Van Essen & Deyoe, 1995; Zeki, 1992, 1993, for reviews). Cognitive psychologists are in the business of carrying out functional analyses, insofar as they proffer box-and-arrow models as explanations of the capacities of humans, where those models specify functionally defined components of the brain's information-processing machinery. Consider, for example, the human ability to shadow speech (i.e., the ability to repeat immediately what is spoken to one). Cognitive psychologists have proposed a rudimentary functional analysis of this system-level capacity according to which there are three routes from auditory input to speech output (see Ellis & Young, 1988; and Figure 3). Each route consists of a small series of distinct information-processing operations. One route, for example ('Route 2' in Figure 3), consists in the following four operations: (a) the extraction of individual speech sounds (e.g., phonemes) from the sound wave, (b) the registration of familiar spoken words, (c) the conversion of the representations of the auditory form of words to representations of the spoken form of words, and (d) the segmentation of the representations of the spoken form of whole words into representations of individual speech sounds (e.g., phonemes). The general proposal is that the system-level capacity of shadowing speech is realized in virtue of the performance of the information- processing functions specified in any one of the three routes. It is assumed that the distinction between these three routes will be reflected in the neural substrate, but that, for the purposes of the functional analysis, it does not matter exactly what that substrate is. Moreover, the functionally defined components produced by a functional analysis need not match up, one-to-one, with the anatomically defined components produced by a structural analysis. The positing of functionally defined components is not meant to imply spatial localization; the 'components' featuring in such accounts are not geographical units, but rather, are units defined solely by some function. Of course it might be the case that some functional components map neatly onto individual geographical components or locations, but this need not be so for all cases. It is entirely possible for a distinct functional component to be implemented by widely dispersed (and possibly diverse) physical components. Conversely, it is possible for a single physical component to implement more than one distinct function. To take some simple examples: birds' wings can act as thermo-regulators as well as instruments of flight, the human tongue has roles in eating, tasting and talking, and a knife can be used for both cutting and spreading. And in a system as complex as the brain, there is likely to be much "multiple, superimposed functionality", as Dennett (1991, p.273) puts it. That is, it is likely to be the case that the mapping of functions to neural structures will be a complex, multi-layered affair.
So functional analysis is to be clearly distinguished from structural analysis. Functional analysis, like structural analysis, involves parsing a system into constituent mechanisms. In structural analysis, mechanisms are individuated along the boundaries of physically defined units (e.g., anatomically distinct neural circuits), whereas in functional analysis, mechanisms are individuated along boundaries that distinguish individual processes or functions. Functional analysis is also to be clearly distinguished from capacity analysis. Functional analyses are accounts of processes, of things done and of the order in which they are done, and so bear some resemblance to capacity analyses. But whereas capacity analysis is concerned with what it is that whole systems do or can do, functional analysis is concerned with the things that parts of whole systems do that enable those whole systems to have certain capacities and properties. In short, capacity analysis is concerned with the dispositions of whole systems, whereas functional analysis is concerned with the bases of those dispositions. (Structural analysis is also often concerned with the bases of dispositions.) 4. A MISTAKE TO AVOID, AND A POSSIBLE LIMITATION To recapitulate: Given that an organism (or more generally, a whole system) has a certain property or capacity, two questions that one can ask of it are these: (1) What other properties or capacities does the whole organism have in virtue of which it has the first property or capacity? (2) What is the material and/or functional composition of the whole organism in virtue of which it has the specified property or capacity; that is, roughly, what structures and processes inside the organism enable that whole organism to do the things that it can do? Questions of type (1) can be answered by conducting capacity analyses, while questions of type (2) can be answered by conducting componential analyses (i.e., functional and structural analyses). When we are doing decompositional analysis, especially functional analysis, there is a general error that we need to take particular care to avoid, but which is not a mistake that could be pinned to particular claims about subsystems. The error is simply assuming, without evidence, that there is a box for every capacity of a whole system. In other words, it is a mistake to assume that, for any capacity of a whole system, one can posit a corresponding information-processing subsystem that purportedly underpins that capacity. We can perceive espresso machines, coffee cups and cakes, for example, but there is abundant evidence showing that we do not have separate espresso machine, coffee cup and cake perceptual systems. [6] Capacity analyses are neutral with respect to mechanism. As we have seen (Section 3), this is a point that Cummins (1983) is well aware of. When we conduct a capacity analysis, Cummins says, we do so "without reference to an instantiating system" (p.29). The capacities featuring in the analysis are ascribed to the whole system, not to parts of that system; "indeed, this analysis seems to put no constraints at all on ... componential analysis." (p.30). A given componential analysis will quite probably be consistent with various capacity analyses, and a given capacity analysis will allow for various componential analyses. Thus, the claim that, for a given capacity or property specified in a capacity analysis, there is a corresponding subsystem identified in the componential analysis that subserves that capacity or property, is bold and quite likely false. That is, it is quite probable that the boundaries of many of the capacities of certain whole systems, such as the human organism or its brain, do not line up with the boundaries of the subsystems specified in functional (or structural) analyses of those systems. Is Cummins right to claim that a capacity "analysis seems to put no constraints at all on ... [a] componential analysis"? While I agree that this is true in general, it is nevertheless my view that a certain type of capacity analysis can constrain componential analyses. Although I shall not present the argument here, it is my view that accurate information-processing theories are more easily obtainable, and mistaken theories more easily avoided, if decompositional analyses are motivated by evolutionary considerations. In short, the view is that evolutionary biology is an important, or even crucial, source of guidance and constraint for componential analyses in psychology (see e.g., Atkinson, 1997; Cosmides, Tooby & Barkow, 1992; Cosmides & Tooby, 1994). I finish the paper by briefly considering a possible limitation on the explanatory scope of decompositional analyses in psychology. This limitation centres upon what has become known as the personal/subpersonal distinction, that is, the distinction between the level at which we talk of the individual human being or person, and the level at which we talk of organs and other body parts, especially the brain. This distinction was vividly captured by Dennett (1969, pp.93-94):
"When we have said that a person has a sensation of pain, locates it and is prompted to react in a certain way, we have said all there is to say within the scope of this vocabulary. We can demand further explanation of how a person happens to withdraw his hand from the hot stove, but we cannot demand further explanations in terms of 'mental processes'. Since the introduction of unanalysable mental qualities leads to a premature end to explanation, we may decide that such introduction is wrong, and look for alternative modes of explanation. If we do this we must abandon the explanatory level of people and their sensations and activities and turn to the sub- personal level of brains and events in the nervous system. But when we abandon the personal level in a very real sense we abandon the subject matter of pains as well. When we abandon mental process talk for physical process talk we cannot say that the mental process analysis [my emphasis] of pain is wrong, for our alternative analysis [my emphasis] cannot be an analysis of pain at all, but rather of something else - the motions of human bodies or the organization of the nervous system."
To a first approximation, then, personal-level phenomena are those picked out in the conceptual scheme of commonsense or folk psychology (mental states and processes, in short). Subpersonal-level phenomena, in contrast, are those picked out by the conceptual scheme of the biological sciences, and in particular, by the brain and cognitive sciences. Notice that, in the final sentence of the above passage, Dennett identifies (and, it seems, runs together) two different types of analysis. One of these is conceptual analysis, the other is what I call componential analysis. Conceptual analysis involves philosophical or conceptual definition. The idea is that a concept is constituted by, or defined in terms of, a set of different concepts. For example, the concept of bachelor is defined in terms of the concepts of unmarried and of man. Componential analysis, in contrast, implies material or functional constitution, as we have seen (Section 3). Dennett points out that there is a limit on the scope or depth of the conceptual analysis of mental state types. The thought is that there is only so far one can go with a conceptual analysis of, for example, pain; a point will be reached where all that is said is all that can be said at that level (i.e., the personal level). And from what Dennett says in the above passage, it seems that he feels that there is something more to a complete explanatory story of a personal-level capacity or property than can be provided by a conceptual analysis. The overall explanatory story is to be made more complete by the provision of a componential analysis. The performance of a componential analysis is the performance of a very different type of explanatory project, which requires the abandonment of one level of description and explanation - the personal level - for another, very different level - the subpersonal level. What can subpersonal sciences, such as information-processing psychology, tell us about persons? It is thought by some (e.g., Hornsby, 1997; McDowell, 1994a, 1994b; Sellars, 1963) that our everyday, folk psychological understanding of humans qua persons cannot be assimilated into a scientific understanding of humans qua natural systems. On this view, cognitive psychology and neuroscience may be able to inform the study of persons - namely, by specifying the neurological and information-processing entities and events that enable humans to have the capacities and properties that feature in folk psychological discourse - but no natural science can tell us what it is to be a person. McDowell (1994a) applauds Dennett's distinction between the personal and subpersonal levels, but suggests that the essence of the distinction, as it should be drawn, is one between organisms and their parts, not between persons and non-persons. For McDowell, what marks the distinction between an organism and its parts is that an organism is a system that we describe as "more or less competently inhabiting an environment" (p.196). "In this context," McDowell continues (1994a, p.196):
"we ask questions like the following: what features of the environment would a creature need to become informed of, in order to live in it with precisely the competence that ... [those particular creatures] display?"
Questions of this sort are to be distinguished from questions of a second kind that can be asked about an organism that competently inhabits an environment, namely, questions about the nature and operation of the organism's parts that enable the organism to live competently in an environment. So, for example, what it is that enables an organism to become informed of features of its environment is that organism's perceptual equipment. And the question of how it is that an organism is informed of features of its environment will be answered by accounts of how that perceptual equipment works. Questions of the first kind are examples of what McDowell calls "constitutive" questions. Questions of the second kind are examples of what he calls "enabling condition" questions. As I see this distinction, constitutive questions are to be answered by conceptual analysis, whereas enabling condition questions are to be answered by componential analysis. In advancing this distinction between constitutive and enabling condition accounts, McDowell has given us reason to separate clearly talk of organisms from talk of their parts. On his view, to characterize organisms, including human organisms, as "more or less competently inhabiting an environment" (1994a, p.196) is to characterize them as minded creatures, that is, in folk psychological terms. What McDowell's view amounts to, then, is that the personal or 'organismal' level is set apart from subpersonal or 'suborganismal' levels of description and explanation - that is, from the system and subsystem levels - to the extent that consummate constitutive accounts of these 'higher-level' phenomena do not require the inclusion of facts about, for example, the brain. Indeed, while McDowell admits that system-level and subsystem-level facts about the brain and its evolution may well help shed light upon personal-level phenomena, he nevertheless holds that such facts are irrelevant to constitutive accounts of those phenomena. On his view, enabling condition theories, including cognitive scientific theories, do not tell us what it is to be a minded creature; only constitutive accounts can tell us that. In sum, componential analyses cannot be used to explain truly personal-level phenomena, that is, concepts that feature in conceptual analyses of mental states and processes. There are two distinct types of analysis, and one cannot subsume the other. One might take McDowell's view as a serious challenge to the explanatory scope and power of cognitive psychology. I do not. There is a clear distinction between the sorts of question one can ask at the personal and subpersonal levels, as McDowell is at pains to emphasize. Consider, for example, attempts to explain consciousness. Questions such as, What is it for a person to be conscious, or to be in conscious states?, are clearly personal- level questions. Cognitive psychologists do not attempt to answer questions of this sort. But they do attempt to answer questions concerning the enabling conditions for consciousness; questions such as, What is it in virtue of which a person can be conscious, or be in conscious states? REFERENCES Atkinson, A. P. (1997). Consciousness boxes and cognitive psychological explanation. Unpublished doctoral thesis, Oxford University. Bechtel, W., & Richardson, R. C. (1993). Discovering complexity: Decomposition and localization as strategies in scientific research. Princeton, NJ: Princeton University Press. Bruce, V., & Young, A. (1986). Understanding face recognition. British Journal of Psychology, 77, 305-27. Churchland, P. S. (1986). Neurophilosophy: Toward a unified science of the mind/brain. Cambridge, MA: MIT Press/ Bradford Books. Cosmides, L., & Tooby, J. (1994). Beyond intuition and instinct blindness: toward an evolutionary rigorous cognitive science. Cognition, 50, 41-77. Cosmides, L., Tooby, J., & Barkow, J. H. (1992). Introduction: Evolutionary psychology and conceptual integration. In J. H. Barkow, L. Cosmides, and J. Tooby (Eds.), The adapted mind: Evolutionary psychology and the generation of culture (pp.3-15). New York: Oxford University Press. Cowey, A. (1985). Aspects of cortical organization related to selective attention and selective impairments of visual perception: A tutorial review. In M. I. Posner and O. S. M. Marin (Eds.), Attention and Performance, XI. Hillsdale, NJ: Lawrence Erlbaum. Crane, T. (Ed.) (1996). Dispositions: A debate. London and New York: Routledge. Cummins, R. (1983). The nature of psychological explanation. Cambridge, MA: MIT Press/ Bradford Books. Dennett, D. C. (1969). Content and consciousness. London and New York: Routledge. Dennett, D. C. (1978). Toward a cognitive theory of consciousness. In C. W. Savage (Ed.), Minnesota studies in the philosophy of science. Minneapolis: University of Minnesota Press. Dennett, D. C. (1990). The interpretation of texts, people, and other artifacts. Philosophy and Phenomenological Research, 50, 177-194. Dennett, D. C. (1991). Consciousness explained. Boston: Little, Brown & Co. Dennett, D. C. (1994). Cognitive science as reverse engineering: Several meanings of 'top- down' and 'bottom-up'. In D. Prawitz, B. Skyrms, and D. Westerstahl (Eds.), Proceedings of the 9th international congress of logic, methodology and philosophy of science. Amsterdam: North-Holland. Dennett, D. C. (1995). Darwin's dangerous idea. Evolution and the meanings of life. New York: Simon & Schuster. Ellis, A. and Young, A. W. (1988). Human cognitive neuropsychology. Hove, Sussex: Lawrence Erlbaum. Fodor, J. A. (1974). Special sciences (or: The disunity of science as a working hypothesis). Synthese, 28, 97-115. Fodor, J. A. (1989). Making mind matter more. Philosophical Topics, 17, 59-80. Hornsby, J. (1997). Simple mindedness: In defense of naive naturalism in the philosophy of mind. Cambridge, MA and London: Harvard University Press. Knapp, T. J. (1986). The emergence of cognitive psychology in the latter half of the twentieth century. In T. J. Knapp and L. C. Robertson (Eds.), Approaches to cognition: Contrasts and controversies (pp.13-35). Hillsdale, NJ: Lawrence Erlbaum. Levine, J. (1983). Materialism and qualia: the explanatory gap. Pacific Philosophical Quarterly, 64, 354-61. Levine, J. (1993). On leaving out what it's like. In M. Davies and G. W. Humphreys (Eds.), Consciousness: Psychological and philosophical essays. Oxford: Basil Blackwell. Marr, D. (1977). Artificial intelligence - a personal view. Artificial Intelligence, 9, 37-48. Marr, D. (1982). Vision. New York: W. H. Freeman & Co. McDowell, J. (1994a). The content of perceptual experience. The Philosophical Quarterly, 44, 190-205. McDowell, J. (1994b). Mind and world. Cambridge, MA and London: Harvard University Press. Pietroski, P., & Rey, G. (1995). When other things aren't equal: Saving ceteris paribus laws from vacuity. British Journal for the Philosophy of Science, 46, 81-110. Sellars, W. (1963). Philosophy and the scientific image of man. In W. Sellars, Science, perception and reality. London: Routledge & Kegan Paul. Van Essen, D., & Deyoe, E. (1995). Concurrent processing in the primate visual cortex. In M. Gazzaniga (Ed.), The cognitive neurosciences. Cambridge, MA: MIT Press/ Bradford Books. Wimsatt, W. C. (1976). Reductionism, levels of organization, and the mind-body problem. In G. G. Globus. G. Maxwell, and I. Savodnik (Eds.), Consciousness and the brain: A scientific and philosophical inquiry (pp.205-267). New York and London: Plenum Press. Zeki, S. (1992). The visual image in mind and brain. Scientific American, 267 (3), 69-76. Zeki, S. (1993). A vision of the brain. Oxford: Blackwell Scientific Publications. FOOTNOTES 1. This is the extended text of a paper presented at the 6th Annual Conference of the European Society for Philosophy and Psychology, Padua, Italy, 1997. A draft of this paper appeared in Connexions, an electronic journal in cognitive science ( http://www.shef.ac.uk/~phil/connex/journal/atkinson.html ). Earlier versions of this paper were presented at the Joint Session of the Aristotelian Society and the Mind Association, Dublin, 1996; and the Wolfson Philosophy Society, Oxford, 1996. My thanks to all those who commented on these presentations. I am indebted to Matthew Elton, Christoph Hoerl, Neil Manson, Paul Pietroski and Tony Stone for their advice and for many helpful conversations related to the material presented here, and to the wider project from which it is drawn, i.e., my D.Phil. thesis in the University of Oxford. I would like to extend a special thankyou to my supervisor, Martin Davies, for countless hours of tuition, incisive advice and stimulating philosophical discussion. 2. A difference in our views that I shall not be explaining here is this: Both Cummins and I agree that it can be very illuminating to decompose a disposition of a system into simpler dispositions, and in so doing to hold off questions about the material substrate of that system. However, whereas Cummins eschews all reference to teleology in his conception of 'function', I accept a 'teleo-evolutionary' conception, and argue that cognitive psychological theories should be grounded upon this notion of 'function' rather than Cummins' more austere notion (see Atkinson, 1997). 3. A potentially fatal dilemma faces this standard view of ceteris paribus ('cp') laws: either such laws are merely vacuous, or, if all the 'other things being equal' clauses are filled in, then the laws become extremely complex and unwieldy, with little or no application to conditions in the real world. Pietroski and Rey (1995) attempt to surmount this dilemma by way of an advocacy of a somewhat different view of cp-laws. On the more usual reading, when a cp-law is found not to apply, one has to drop to a more basic level science to explain why so. On Pietroski and Rey's view, in contrast, one does not need to drop to a lower level of description and explanation for an explanation of why a cp-law did not apply on a certain occasion. Exceptions to such laws "are to be explained as the result of interference from independent systems" (p.87), that is, roughly, as the result of interactions of one or more of the systems specified or implied in the cp-law with one or more systems not covered by the law (where both sets of systems are specified at the same level). Pietroski and Rey (1995, p.91): "On our view, ... a chemist holding that cp(PV = nRT), is committed to the following: if a gas sample G is such that PV ? nRT, there are independent factors (e.g., electrical attraction) that explain why PV ? nRT with respect to G." 4. This way of putting the point is merely for ease of expression. Of course there might well be further neural circuits intermediate between a 'small scale' neural circuit and the whole brain; it all really depends on how one carves up the brain's - or for that matter, any given system's - subsystems. 5. It will become apparent that there is much overlap between the tri-level view of cognitive psychological explanation as I shall present it (viz., capacity analysis, functional analysis and structural analysis) and Marr's (1977, 1982) scheme of three levels of analysis (viz., the "computational" or ecological, "algorithmic" and implementational levels). There is one important difference between Marr's scheme and mine, however, which is that I emphasize the shift from the system level to the subsystem level whereas Marr does not. 6. Notice that this is a more general error than the familiar one discussed by, for example, Churchland (1986), viz., the mistake of accepting, as a general inferential scheme, the following: a particular brain region, A, is the 'centre' for some capacity, C, since (1) A is lesioned in patient Y, and (2) Y no longer has C. In any given case, such an inference might be more or less true (perhaps more correctly expressed as 'A underpins or subserves, or has elements that underpin or subserve, C'). But its general application "will yield a bizarre catalogue of centers - including, for example, a center for inhibiting religious fanaticism, since lesions in certain areas of the temporal lobe sometimes result in a patient's acquiring a besotted religious zeal. Add to this centers for being able to make gestures on command, for prevention of halting speech, for inhibition of cursing and swearing, and so on ..." (Churchland, 1986, p.164). © Anthony Atkinson