---
abstract: |-
The Turing Test is just a methodological constraint forcing us to scale up to an organisms' full functional
capacity. This is still just an epistemic matter, not an ontic one. Even a candidate in which we have successfully
reverse-engineered all human capacities is not guaranteed to have a mind. The right level of convergence,
however, is total robotic capacity; symbolic capacity alone (the standard Turing Test) is underdetermined,
whereas full neurosimilitude is overdetermined.
altloc:
- http://www.cogsci.soton.ac.uk/cgi/psyc/newpsy?11.078
chapter: ~
commentary: ~
commref: 'Green (2000) Is AI the Right Method for Cognitive Science?, PSYCOLOQUY 11(061)'
confdates: ~
conference: ~
confloc: ~
contact_email: ~
creators_id: []
creators_name:
- family: Harnad
given: Stevan
honourific: ''
lineage: ''
date: 2000
date_type: published
datestamp: 2001-06-26
department: ~
dir: disk0/00/00/16/49
edit_lock_since: ~
edit_lock_until: ~
edit_lock_user: ~
editors_id: []
editors_name: []
eprint_status: archive
eprintid: 1649
fileinfo: /style/images/fileicons/text_html.png;/1649/1/psyc.00.11.078.ai%2Dcognitive%2Dscience.18.harnad
full_text_status: public
importid: ~
institution: ~
isbn: ~
ispublished: pub
issn: ~
item_issues_comment: []
item_issues_count: 0
item_issues_description: []
item_issues_id: []
item_issues_reported_by: []
item_issues_resolved_by: []
item_issues_status: []
item_issues_timestamp: []
item_issues_type: []
keywords: |-
artificial intelligence, behaviorism, cognitive science, computationalism, Fodor, functionalism, Searle, Turing Machine,
Turing Test.
lastmod: 2011-03-11 08:54:43
latitude: ~
longitude: ~
metadata_visibility: show
note: ~
number: 78
pagerange: ~
pubdom: FALSE
publication: Psycoloquy
publisher: ~
refereed: TRUE
referencetext: |-
Chiappe, D.L. & Kukla, A. (2000) Artificial Intelligence and Scientific Understanding. PSYCOLOQUY 11(064)
ftp://ftp.princeton.edu/pub/harnad/Psycoloquy/2000.volume.11/ psyc.00.11.064.ai-cognitive-science.4.chiappe
http://www.cogsci.soton.ac.uk/cgi/psyc/newpsy?11.064
Chiappe, D.L. & Kukla, A. (1993) Artificial Intelligence and Scientific understanding. Cognoscenti 1: 7-9.
Dennett, D.C. (1993) Discussion (passim) In: Bock, G.R. & Marsh, J. (Eds.) Experimental and Theoretical Studies of
Consciousness. CIBA Foundation Symposium 174. Chichester: Wiley
Dennett, D.C. (1994) Cognitive Science as Reverse Engineering: Several Meanings of "Top Down" and "Bottom Up". In:
Prawitz, D., & Westerstahl, D. (Eds.) International Congress of Logic, Methodology and Philosophy of Science.
Dordrecht: Kluwer International Congress of Logic, Methodology, and Philosophy of Science (9th: 1991)
http://cogsci.soton.ac.uk/~harnad/Papers/Py104/dennett.eng.html
Fodor, J.A. (1981) The Mind-Body Problem. Scientific American 244: 114-23.
Fodor, J.A. (1983) The Modularity of Mind. Cambridge MA: MIT Press.
Fodor, J.A. (1991) Replies. In B. Loewer & G. Rey (Eds.) Meaning in Mind: Fodor and his Critics (pp. 255-319).
Cambridge MA: Blackwell.
Green, C.D. (2000a) Is AI the Right Method for Cognitive Science? PSYCOLOQUY 11(061)
ftp://ftp.princeton.edu/pub/harnad/Psycoloquy/2000.volume.11/ psyc.00.11.061.ai-cognitive-science.1.green
http://www.cogsci.soton.ac.uk/cgi/psyc/newpsy?11.061
Green, C.D. (1993a) Is AI the Right Method For Cognitive Science? Cognoscenti 1: 1-5
Green, C.D. (2000b) Empirical Science and Conceptual Analysis Go Hand in Hand. PSYCOLOQUY 11(071)
ftp://ftp.princeton.edu/pub/harnad/Psycoloquy/2000.volume.11/ psycoloquy.00.11.071.ai-cognitive-science.11.green
http://www.cogsci.soton.ac.uk/cgi/psyc/newpsy?11.071
Green, C.D. (1993b) Ontology Rules! (But not Absolutely). Cognoscenti 1: 21-28.
Harnad, S. (1982) Consciousness: An afterthought. Cognition and Brain Theory 5: 29 - 47.
http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad82.consciousness.html
Harnad, S. (ed.) (1987) Categorical Perception: The Groundwork of Cognition. New York: Cambridge University Press.
http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad87.categorization.html
Harnad, S. (1989) Minds, Machines and Searle. Journal of Theoretical and Experimental Artificial Intelligence 1: 5-25.
http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad89.searle.html
Harnad, S. (1990) The Symbol Grounding Problem. Physica D 42: 335-346.
http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad90.sgproblem.html
Harnad, S. (1991) Other bodies, Other minds: A machine incarnation of an old philosophical problem. Minds and Machines
1: 43-54. http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad91.otherminds.html
Harnad, S. (1992a) Connecting Object to Symbol in Modeling Cognition. In: A. Clarke and R. Lutz (Eds.) Connectionism
in Context Springer Verlag. http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad92.symbol.object.html
Harnad, S. (1992b) The Turing Test Is Not A Trick: Turing Indistinguishability Is A Scientific Criterion. SIGART Bulletin
3(4) (October) 9 - 10. http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad92.turing.html
Harnad, S. (1993a) Grounding Symbols in the Analog World with Neural Nets. Think 2(1) 12 - 78 (Special issue on
"Connectionism versus Symbolism," D.M.W. Powers & P.A. Flach, eds.).
http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad93.symb.anal.net.html
http://cwis.kub.nl/~fdl/research/ti/docs/think/2-1/index.stm
Harnad, S. (1993b) Problems, Problems: The Frame Problem as a Symptom of the Symbol Grounding Problem.
PSYCOLOQUY 4(34) frame-problem.11
http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad93.frameproblem.html
Harnad, S. (1993c) Symbol Grounding is an Empirical Problem: Neural Nets are Just a Candidate Component. Proceedings
of the Fifteenth Annual Meeting of the Cognitive Science Society. NJ: Erlbaum
Harnad, S. (1994) Levels of Functional Equivalence in Reverse Bioengineering: The Darwinian Turing Test for Artificial
Life. Artificial Life 1(3): 293-301. Reprinted in: C.G. Langton (Ed.). Artificial Life: An Overview. MIT Press 1995.
http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad94.artlife2.html
Harnad, S. (1995a) Grounding Symbolic Capacity in Robotic Capacity. In: Steels, L. and R. Brooks (eds.) The Artificial
Life Route to Artificial Intelligence: Building Embodied Situated Agents. New Haven: Lawrence Erlbaum. Pp. 277-286.
http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad95.robot.html
Harnad, S, (1995b) Does the Mind Piggy-Back on Robotic and Symbolic Capacity? In: H. Morowitz (ed.) "The Mind, the
Brain, and Complex Adaptive Systems." Santa Fe Institute Studies in the Sciences of Complexity. Volume XXII. P.
204-220. http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad95.mind.robot.html
Harnad, S. (2000) Turing Indistinguishability and the Blind Watchmaker. In: Mulhauser, G. (ed.) "Evolving Consciousness"
Amsterdam: John Benjamins (in press) http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad98.turing.evol.html
Harnad, S. (2001) Minds, Machines, and Turing: The Indistinguishability of Indistinguishables. Journal of Logic, Language,
and Information (JoLLI) special issue on "Alan Turing and Artificial Intelligence" (in press)
http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad00.turing.html
Hayes, P., Harnad, S., Perlis, D. & Block, N. (1992) Virtual Symposium on Virtual Mind. Minds and Machines 2:
217-238. http://www.cogsci.soton.ac.uk/~harnad/Papers/Harnad/harnad92.virtualmind.html
Plate, T. (2000) Caution: Philosophers at work. PSYCOLOQUY 11(70)
ftp://ftp.princeton.edu/pub/harnad/Psycoloquy/2000.volume.11/ psyc.00.11.070.ai-cognitive-science.10.plate
http://www.cogsci.soton.ac.uk/cgi/psyc/newpsy?11.070
Plate, T. (1993) Reply to Green. Cognoscenti 1: 13.
Searle, J. R. (1980) Minds, Brains and Programs. Behavioral and Brain Sciences 3: 417-424.
http://www.cogsci.soton.ac.uk/bbs/Archive/bbs.searle2.html
Zelazo, P.D. (2000) The nature (and artifice) of cognition. PSYCOLOQUY 11(076)
ftp://ftp.princeton.edu/pub/harnad/Psycoloquy/2000.volume.11/ psyc.00.11.076.ai-cognitive-science.16.zelazo
http://www.cogsci.soton.ac.uk/cgi/psyc/newpsy?11.076
Zelazo, P.D. (1993) The Nature (and Artifice) of Cognition. Cognoscenti 1: 18-20
relation_type: []
relation_uri: []
reportno: ~
rev_number: 8
series: ~
source: ~
status_changed: 2007-09-12 16:39:26
subjects:
- cog-psy
succeeds: ~
suggestions: ~
sword_depositor: ~
sword_slug: ~
thesistype: ~
title: 'The Convergence Argument in Mind-Modelling: Scaling Up from Toyland to the Total Turing Test'
type: journale
userid: 63
volume: 11