Skip to main content

2018 | OriginalPaper | Buchkapitel

Where Intelligence Lies: Externalist and Sociolinguistic Perspectives on the Turing Test and AI

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Turing’s Imitation Game (1950) is usually understood to be a test for machines’ intelligence; I offer an alternative interpretation. Turing, I argue, held an externalist-like view of intelligence, according to which an entity’s being intelligent is dependent not just on its functions and internal structure, but also on the way it is perceived by society. He conditioned the determination that a machine is intelligent upon two criteria: one technological and one sociolinguistic. The Technological Criterion requires that the machine’s structure enables it to imitate the human brain so well that it displays intelligent-like behavior; the Imitation Game tests if this Technological Criterion was fulfilled. The Sociolinguistic Criterion requires that the machine be perceived by society as a potentially intelligent entity. Turing recognized that in his day, this Sociolinguistic Criterion could not be fulfilled due to humans’ chauvinistic prejudice towards machines; but he believed that future development of machines displaying intelligent-like behavior would cause this chauvinistic attitude to change. I conclude by discussing some implications Turing’s view may have in the fields of AI development and ethics.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
My reading is supported by several pieces of non-orthodox commentaries of Turing scattered throughout the literature, such as Whitby (1996), Boden (2006, pp. 1346–1356), Sloman (2013), and especially Proudfoot (2005; 2013).
Some of the arguments suggested in this paper have appeared in Danziger (2016).
 
2
As Piccinini (2000), Proudfoot (2013), and others have pointed out, Turing uses the terms “thought” and “intelligence” interchangeably. Although I will not differentiate between the terms, for reasons of uniformity I shall usually use the term “intelligence”.
 
3
Almost all commentaries on – and attacks against – Turing’s paper can be classified into one of the two streams of interpretation described in Sects. 2.1 and 2.2; see Proudfoot (2013) for detailed analysis and critique of these interpretations. Other ways of classification can be found in Saygin et al. (2000) and in Oppy and Dowe (2011).
 
4
Also Searle’s interpretation of Turing is behavioristic: “The Turing Test is typical of the tradition in being unashamedly behavioristic and operationalistic” (Searle 1980, p. 423; cf. next footnote). References to other behavioristic interpretations can be found in Proudfoot (2013), Copeland (2004, pp. 434–435), and Moor (2001, pp. 81–82).
 
5
This is the crux of perhaps the two most well-known arguments against the IG, namely, Searle’s Chinese Room (Searle 1980) and Block’s Blockhead / Aunt Bubbles Machine (Block 1981; 1995): Intelligence, they maintain, cannot be captured in behavioral terms alone. (Note that both arguments belong to the behavioristic school of interpretation, in that they assume that the IG is intended to be a behavioral test for intelligence.)
 
6
The main proponent of the school of inductive interpretations is Moor (1976; 2001). Other inductive interpretations can be found in Watt (1996) and Schweizer (1998).
 
7
The ideas in this section draw partly on Proudfoot (2005; 2013). As I shall show later, my interpretation of Turing differs from Proudfoot’s in small but crucial points; to prevent inaccuracies I shall refrain for now from mentioning her take on the subjects discussed, despite my great debt to her work.
 
8
Turing’s approach bears resemblance also to Dennett’s “intentional stance” (Dennett 1987a).
 
9
In Sects. 3.2 and 4.3 I shall bring further textual evidence for this being Turing’s approach, and shall briefly discuss what might have motivated Turing into adopting such a stance.
 
10
Turing is trying to prove that the existence of an intelligent machine is possible, and is not merely asking if it possible. Therefore he will try to show that machines fulfill a sufficient condition for being (perceived as) intelligent, and will put less emphasis on the necessary conditions.
 
11
There is no need to point out here which properties of the “learning machine” are necessary conditions for perceiving a system as intelligent; all that is being claimed is that a “learning machine” indeed has these properties, whatever they may be.
 
12
It will later become clear why this claim is labeled “minor”.
 
13
Hodges (2014, p. 530) explains in a similar way the difference between Turing’s 1948 and 1950 papers.
 
14
“Intelligent-like behavior” may be roughly defined as “behavior that under regular circumstances cannot be differentiated from that of a human”.
 
15
The Technological Criterion (1950) is closely connected to the Minor-Technological Claim (1947, 1948) but is more “demanding” (as explained above, Sect. 4.1); that is why the 1947–1948 claim is labeled “minor”.
 
16
Bringsjord et al. (2001) mention a similar idea of “restricted epistemic relation”: They suggest the “Lovelace Test” for intelligence in which “not knowing how a system works” is a necessary condition for attributing intelligence to it. The fundamental difference between the Lovelace Test and Turing’s IG will be explained later (fn. 23).
 
17
Other clear remarks of Wittgenstein in this spirit are Wittgenstein (2009, §281) and Wittgenstein (1958, p. 47). The similarity between Turing’s and Wittgenstein’s ideas here has been pointed out also by Boden (2006, p. 1351) and Chomsky (2008, p. 104).
 
18
The Sociolinguistic Criterion (1950) is closely connected to the Sociological Claim (1947, 1948) mentioned in Sect. 3.2. The addition of the “linguistic” component will soon be explained.
 
19
At this point one might raise the following objection: “Your reading boldly ignores the next sentence in Turing’s paper, in which he supposedly predicts that in fifty years there would be intelligent machines (1950, p. 442): ‘Nevertheless I believe that at the end of the century the use of words and general educated opinion will have altered so much that one will be able to speak of machines thinking without expecting to be contradicted.’ This implies that Turing identified success in constructing machines that do well in the Game – with success in creating intelligent machines; the timeframe in both sentences is the same (the year 2000), and so they seem to be referring to the same futuristic occurrence!” My reply, in short, is that this objection is based on an incorrect – albeit very common – reading of the passage in Turing’s paper. Turing, I claim, makes two different predictions here, and these predictions are connected causally but not logically. “Doing well in the IG” is not the same as “being intelligent”. The IG, I insist, is not a test for intelligence, but a test only for the Technological Criterion of intelligence: it tests if a system’s behavior is intelligent-like. (I shall return to this issue in Sect. 5.1.)
 
20
Aaron Sloman, too, sees the IG as Turing’s way of defining a technological challenge, and not as a test for intelligence (Sloman 2013). In an earlier version of his paper Sloman expresses his dissatisfaction with the orthodox interpretations of the IG; I found myself wholly identifying with his words (my italics): “It is widely believed that Turing proposed a test for intelligence. This is false. He was far too intelligent to do any such thing, as should be clear to anyone who has read his paper…”
 
21
To develop this point further: A real test for a system’s intelligence would check if the system is perceived as intelligent by society as a whole, in an ongoing manner, in normal life situations. But if that were to happen there would be no need for an intelligence test, because “society perceiving a system as intelligent” is the definition of a system’s being intelligent, not a sign of it! (See the Social-Relationist Premise, Sect. 3.1.)
 
22
This is how Turing’s prediction was understood by Mays (1952, pp. 149–151), Beran (2014) and others. (Piccinini 2000 understands that Turing hopes such a change will occur.) For an illuminating discussion regarding the possibility of this sort of change (not concerning Turing’s paper) see Torrance (2014).
 
23
The main difference between Turing’s IG and Bringsjord et al.’s “Lovelace Test” mentioned above (fn. 16) is that while the IG is descriptive, the Lovelace Test is normative (see Bringsjord et al. 2001, p. 9).
 
24
Sloman makes a similar point and says that while computers are now doing much cleverer things, “increasing numbers of humans have been learning about what computers can and cannot do” (Sloman 2013, p. 3). Indeed, getting humans to attribute intelligence to machines might become harder with time.
 
25
In his brief reply to the “Argument from Consciousness”, Turing seems to claim that if a machine did well in the IG it would be perceived as conscious too (1950, pp. 445–447; see Michie 1993, pp. 4–7. But cf. Copeland 2004, pp. 566–567). I am of the opinion that likewise intelligence, also consciousness and other mental phenomena can be explained in terms of being perceived by society; I plan to discuss this elsewhere.
 
26
Both properties mentioned were suggested by Mays (1952), in his analysis of Turing’s 1950 paper. Interestingly, Turing himself seems to have viewed both properties as insignificant for intelligence attribution (see Turing 1950, p. 434; Davidson 1990). For a list of other properties that might shape humans’ attitude towards machines, see Torrance (2014).
 
27
Discussions regarding the active role of humans in drawing the borders of the “Charmed Circle” of consciousness or intelligence (relevant also to the issue of animal consciousness and to disputes regarding humans’ attitude towards animals) can be found in Dennett (1987b) and Michie (1993).
 
Literatur
Zurück zum Zitat Beran, O.: Wittgensteinian perspectives on the Turing test. Studia Philosophica Estonica 7(1), 35–57 (2014)CrossRef Beran, O.: Wittgensteinian perspectives on the Turing test. Studia Philosophica Estonica 7(1), 35–57 (2014)CrossRef
Zurück zum Zitat Block, N.: Psychologism and behaviorism. Philos. Rev. 90, 5–43 (1981)CrossRef Block, N.: Psychologism and behaviorism. Philos. Rev. 90, 5–43 (1981)CrossRef
Zurück zum Zitat Block, N.: The mind as the software of the brain. In: Smith, E.E., Osherson, D.N. (eds.) Thinking, pp. 377–425. MIT Press, Cambridge (1995) Block, N.: The mind as the software of the brain. In: Smith, E.E., Osherson, D.N. (eds.) Thinking, pp. 377–425. MIT Press, Cambridge (1995)
Zurück zum Zitat Boden, M.A.: Mind as Machine: A History of Cognitive Science. Oxford University Press, Oxford (2006) Boden, M.A.: Mind as Machine: A History of Cognitive Science. Oxford University Press, Oxford (2006)
Zurück zum Zitat Bringsjord, S., Bello, P., Ferrucci, D.: Creativity, the Turing test, and the (better) Lovelace test. Mind. Mach. 11, 3–27 (2001)CrossRef Bringsjord, S., Bello, P., Ferrucci, D.: Creativity, the Turing test, and the (better) Lovelace test. Mind. Mach. 11, 3–27 (2001)CrossRef
Zurück zum Zitat Chomsky, N.: Turing on the “imitation game”. In: Epstein, R., Roberts, G., Beber, G. (eds.) Parsing the Turing Test, pp. 103–106. Springer, New York (2008) Chomsky, N.: Turing on the “imitation game”. In: Epstein, R., Roberts, G., Beber, G. (eds.) Parsing the Turing Test, pp. 103–106. Springer, New York (2008)
Zurück zum Zitat Coeckelbergh, M.: Robot rights? Towards a social-relational justification of moral consideration. Ethics Inf. Technol. 12(3), 209–221 (2010)CrossRef Coeckelbergh, M.: Robot rights? Towards a social-relational justification of moral consideration. Ethics Inf. Technol. 12(3), 209–221 (2010)CrossRef
Zurück zum Zitat Copeland, B.J. (ed.): The Essential Turing. Oxford University Press, Oxford (2004)MATH Copeland, B.J. (ed.): The Essential Turing. Oxford University Press, Oxford (2004)MATH
Zurück zum Zitat Danziger, S.: Can computers be thought of as thinkers? Externalist and linguistic perspectives on the Turing test. MA thesis, Hebrew University of Jerusalem (2016). [Hebrew] Danziger, S.: Can computers be thought of as thinkers? Externalist and linguistic perspectives on the Turing test. MA thesis, Hebrew University of Jerusalem (2016). [Hebrew]
Zurück zum Zitat Davidson, D.: Turing’s test. In: Said, K., Newton-Smith, W., Viale, R. Wilkes, K. (eds.) Modelling the Mind, pp. 1–12. Clarendon Press, Oxford (1990) Davidson, D.: Turing’s test. In: Said, K., Newton-Smith, W., Viale, R. Wilkes, K. (eds.) Modelling the Mind, pp. 1–12. Clarendon Press, Oxford (1990)
Zurück zum Zitat Dennett, D.C.: The Intentional Stance. MIT Press, Cambridge (1987a) Dennett, D.C.: The Intentional Stance. MIT Press, Cambridge (1987a)
Zurück zum Zitat Dennett, D.C.: Consciousness. In: Gregory, R.L., Zangwill, O.L. (eds.) The Oxford Companion to the Mind, pp. 160–164. Oxford University Press, Oxford (1987b) Dennett, D.C.: Consciousness. In: Gregory, R.L., Zangwill, O.L. (eds.) The Oxford Companion to the Mind, pp. 160–164. Oxford University Press, Oxford (1987b)
Zurück zum Zitat Hodges, A.: Alan Turing: The Enigma. Princeton University Press, Princeton and Oxford (2014)CrossRef Hodges, A.: Alan Turing: The Enigma. Princeton University Press, Princeton and Oxford (2014)CrossRef
Zurück zum Zitat Mallery, J.C.: Thinking about foreign policy: finding an appropriate role for artificially intelligent computers. The 1988 Annual Meeting of the International Studies Association, St. Louis (1988). 10.1.1.50.3333 Mallery, J.C.: Thinking about foreign policy: finding an appropriate role for artificially intelligent computers. The 1988 Annual Meeting of the International Studies Association, St. Louis (1988). 10.1.1.50.3333
Zurück zum Zitat Moor, J.H.: An analysis of the Turing test. Philos. Stud. 30, 249–257 (1976)CrossRef Moor, J.H.: An analysis of the Turing test. Philos. Stud. 30, 249–257 (1976)CrossRef
Zurück zum Zitat Moor, J.H.: The status and future of the Turing test. Mind. Mach. 11, 77–93 (2001)CrossRef Moor, J.H.: The status and future of the Turing test. Mind. Mach. 11, 77–93 (2001)CrossRef
Zurück zum Zitat Piccinini, G.: Turing’s rules for the imitation game. Mind. Mach. 10, 573–582 (2000)CrossRef Piccinini, G.: Turing’s rules for the imitation game. Mind. Mach. 10, 573–582 (2000)CrossRef
Zurück zum Zitat Proudfoot, D.: Rethinking Turing’s test. J. Philos. 110(7), 391–411 (2013)CrossRef Proudfoot, D.: Rethinking Turing’s test. J. Philos. 110(7), 391–411 (2013)CrossRef
Zurück zum Zitat Saygin, A., Cicekli, I., Akman, V.: Turing test: 50 years later. Minds Mach. 10, 463–518 (2000)CrossRef Saygin, A., Cicekli, I., Akman, V.: Turing test: 50 years later. Minds Mach. 10, 463–518 (2000)CrossRef
Zurück zum Zitat Schweizer, P.: The truly total Turing test. Mind. Mach. 8, 263–272 (1998)CrossRef Schweizer, P.: The truly total Turing test. Mind. Mach. 8, 263–272 (1998)CrossRef
Zurück zum Zitat Searle, J.R.: Minds, brains, and programs. Behav. Brain Sci. 3, 417–424 (1980)CrossRef Searle, J.R.: Minds, brains, and programs. Behav. Brain Sci. 3, 417–424 (1980)CrossRef
Zurück zum Zitat Sloman, A.: The mythical Turing test. In: Cooper, S.B., Van Leeuwen, J. (eds.) Alan Turing: His Work and Impact, pp. 606–611. Elsevier, Amsterdam (2013) Sloman, A.: The mythical Turing test. In: Cooper, S.B., Van Leeuwen, J. (eds.) Alan Turing: His Work and Impact, pp. 606–611. Elsevier, Amsterdam (2013)
Zurück zum Zitat Torrance, S.: Artificial consciousness and artificial ethics: between realism and social relationism. Philos. Technol. 27(1), 9–29 (2014)CrossRef Torrance, S.: Artificial consciousness and artificial ethics: between realism and social relationism. Philos. Technol. 27(1), 9–29 (2014)CrossRef
Zurück zum Zitat Turing, A.M.: On computable numbers, with an application to the Entscheidungsproblem. Reprinted in Copeland (2004), pp. 58–90 (1936) Turing, A.M.: On computable numbers, with an application to the Entscheidungsproblem. Reprinted in Copeland (2004), pp. 58–90 (1936)
Zurück zum Zitat Turing, A.M.: Lecture on the automatic computing engine. Reprinted in Copeland (2004), pp. 378–394 (1947) Turing, A.M.: Lecture on the automatic computing engine. Reprinted in Copeland (2004), pp. 378–394 (1947)
Zurück zum Zitat Turing, A.M.: Intelligent machinery. Reprinted in Copeland (2004), pp. 410–432 (1948) Turing, A.M.: Intelligent machinery. Reprinted in Copeland (2004), pp. 410–432 (1948)
Zurück zum Zitat Turing, A.M., Braithwaite, R., Jefferson, G., Newman, M.: Can automatic calculating machines be said to think? Reprinted in Copeland (2004), pp. 494–506 (1952) Turing, A.M., Braithwaite, R., Jefferson, G., Newman, M.: Can automatic calculating machines be said to think? Reprinted in Copeland (2004), pp. 494–506 (1952)
Zurück zum Zitat Whitby, B.: The Turing test: AI’s biggest blind alley? In: Millican, P., Clark, A. (eds.) Machines and Thought: The Legacy of Alan Turing, pp. 53–62. Calderon Press, Oxford (1996) Whitby, B.: The Turing test: AI’s biggest blind alley? In: Millican, P., Clark, A. (eds.) Machines and Thought: The Legacy of Alan Turing, pp. 53–62. Calderon Press, Oxford (1996)
Zurück zum Zitat Wittgenstein, L.: The Blue and Brown Books: Preliminary Studies for the “Philosophical Investigations”. Harper & Row, New York (1958) Wittgenstein, L.: The Blue and Brown Books: Preliminary Studies for the “Philosophical Investigations”. Harper & Row, New York (1958)
Zurück zum Zitat Wittgenstein, L.: Philosophical Investigations (Trans: G.E.M. Anscombe, P.M.S. Hacker, & J. Schulte; revised fourth edition by P.M.S. Hacker & J. Schulte). Wiley-Blackwell, Chichester (2009) Wittgenstein, L.: Philosophical Investigations (Trans: G.E.M. Anscombe, P.M.S. Hacker, & J. Schulte; revised fourth edition by P.M.S. Hacker & J. Schulte). Wiley-Blackwell, Chichester (2009)
Metadaten
Titel
Where Intelligence Lies: Externalist and Sociolinguistic Perspectives on the Turing Test and AI
verfasst von
Shlomo Danziger
Copyright-Jahr
2018
DOI
https://doi.org/10.1007/978-3-319-96448-5_15