Skip to main content

2018 | OriginalPaper | Buchkapitel

Do Machine-Learning Machines Learn?

verfasst von : Selmer Bringsjord, Naveen Sundar Govindarajulu, Shreya Banerjee, John Hummel

Erschienen in: Philosophy and Theory of Artificial Intelligence 2017

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

We answer the present paper’s title in the negative. We begin by introducing and characterizing “real learning” (\(\mathcal {RL}\)) in the formal sciences, a phenomenon that has been firmly in place in homes and schools since at least Euclid. The defense of our negative answer pivots on an integration of reductio and proof by cases, and constitutes a general method for showing that any contemporary form of machine learning (ML) isn’t real learning. Along the way, we canvass the many different conceptions of “learning” in not only AI, but psychology and its allied disciplines; none of these conceptions (with one exception arising from the view of cognitive development espoused by Piaget), aligns with real learning. We explain in this context by four steps how to broadly characterize and arrive at a focus on \(\mathcal {RL}\).

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
The need for the qualifications (i.e. determinate, non-question-begging) should be obvious. The answer to the present paper’s title that a machine which machine-learns by definition learns, since ‘learn’ appears in ‘machine-learn,’ assumes at the outset that what is called ‘machine learning’ today is real learning—but that’s precisely what’s under question; hence the petitio.
 
2
All mathematical models of learning relevant to the present discussion that we are aware of take learning to consist fundamentally in the learning of number-theoretic functions from \(\mathbb {N} \times \mathbb {N} \times \cdots \times \mathbb {N}\) to \(\mathbb {N}\). Even when computational learning was firmly and exclusively rooted in classical recursion theory, and dedicated statistical formalisms were nowhere to be found, the target of learning was a function of this kind; see e.g. (Gold 1965; Putnam 1965), a modern, comprehensive version of which is given in (Jain et al. 1999). We have been surprised to hear that some in our audience aren’t aware of the basic, uncontroversial fact, readily appreciated by consulting the standard textbooks we cite here and below, that machine learning in its many guises takes the target of learning to be number-theoretic functions. A “shortcut” to grasping a priori that all systematic, rigorously described forms of learning in matters and activities computational and mechanistic must be rooted in number-theoretic functions, is to simply note that computer science itself consists in the study and embodiment of number-theoretic functions, defined and ordered in hierarchies (e.g. see Davis and Weyuker 1983). We by the way focus herein on unary functions \(f : \mathbb {N} \mapsto \mathbb {N}\) only for ease of exposition.
 
3
A biconditional isn’t needed. We use only a weaker set of necessary conditions, not a set of necessary and sufficient conditions.
 
4
Not to be be confused with RL, reinforcement learning, in which real learning, as revealed herein, doesn’t happen.
 
5
As many readers will know, Searle’s (1980) Chinese Room Argument (CRA) is intended to show that computing machines can’t understand anything. It’s true that Bringsjord has refined, expanded, and defended CRA (e.g. see Bringsjord 1992, Bringsjord and Noel 2002; Bringsjord 2015), but bringing to bear here this argumentation in support of the present paper’s main claim would instantly demand an enormous amount of additional space. And besides, as we now explain, calling upon this argumentation is unnecessary.
 
6
Since at bottom, as noted (see note 2), the target of learning should be taken for generality and rigor to be a number-theoretic function, it’s natural to consider learning in the realm of the formal sciences.
 
7
Just as (computer) programs can be correct or incorrect, so too proofs can be correct or incorrect. For more on this, see e.g. (Arkoudas and Bringsjord 2007).
 
8
If we regard Turing to have been speaking of modern AI in his famous (Turing, 1950), note then too that his orientation is test-based: he gave here of course the famous ‘Turing Test.’.
 
9
In fact, this is why real learning for humans in mathematics is challenging; see e.g. (Moore 1994).
 
10
Our assumption here thus specifically invokes connectionist ML. But this causes no loss of generality, as we explain by way the “tour” of ML taken in Sect. 6.1, and the fact that the proof in the Appendix, as explained there, is a general method that will work form any contemporary form of ML.
 
11
This is a rough-and-ready extraction from (Jain et al. 1999), and must be sufficient given the space limitations of the present short paper, at least for now. Of course, there are many forms of ML/machine learning in play in AI of today. In Sect. 6.1 we consider different forms of ML in contemporary AI. In Sect. 6.2 we consider different types of “learning” in psychology and allied disciplines.
 
12
Lathrop (1996) shows, it might be asserted, that uncomputable functions can be machine-learned. But in his scheme, there is only a probabilistic approximation of real learning, and—in clear tension with (c1\('\))–(c3)—no proof in support of the notion that anything has been learned. The absence of such proofs is specifically called out in the formal deduction given in the Appendix.
 
13
A pair of additional works help to further seal our case: (Kearns and Vazirani 1994; Shalev- Shwartz and Ben-David 2014). Study of these texts will reveal that \(\mathcal {RL}\) as per (c1\('\))–(c3) is nowhere to be found.
 
14
We of course join epistemological cognoscenti in being aware of Gettier-style cases, but they can be safely left aside here. For the record, Bringsjord claims to have a solution anyway—one that is at least generally in the spirit of Chisholm’s (1966) proposed solution, which involves requiring that the justification in justified-true-belief accounts of knowledge be of a certain sort. For Gettier’s landmark paper, see (Gettier 1963).
 
15
Specifically, we shall see that the formal deduction of the Appendix is actually a method for showing that other forms of “modern” ML, not just those that rely on ANNs, don’t enable machines to really learn anything. E.g., the method can take Bayesian learning in, and yield as output that such learning isn’t real learning.
 
16
Shakespeare himself, or better yet even Ibsen, or better better yet Bellow, couldn’t have invented a story dripping with this much irony—a story in which the machine-learning people persecuted the logicians for building “brittle” systems, and then the persecutors promptly proceeded to blithely build comically brittle systems as their trophies (given to themselves).
 
17
In which is by the way cited hypercomputational artificial neural networks.
 
18
E.g. even beginning textbooks introducing single-variable differential/integral calculus ask for verification of human learning by asking for proofs. The cornerstone and early-on-introduced concept of a limit is accordingly accompanied by requests to students that they supply proofs in order to confirm that they understand this concept. Thus we e.g. have on p. 67 of (Stewart, 2016) a request that our reader prove that
$$\begin{aligned} \lim _{x \rightarrow 3} g(x) = (4x - 5) = 7 \end{aligned}$$
. What machine-learning machine that has learned the function g here can do that?
 
19
This is essentially the Short Short Story Game of (Bringsjord 1998), much harder than such Turing-computable games as Checkers, Chess, and Go, which are all at the same easy level of difficulty (EXPTIME).
 
20
Outside of the present paper, we have carried out a second analysis that confirms this, by examining learning in AI as characterized in (Russell and Norvig 2009), and invite skeptical readers to carry out their own analysis for this textbook, and indeed for any comprehensive, mainstream textbook. The upshot will be the stark fact that \(\mathcal {RL}\), firmly in place since Euclid as what learning in the formal sciences is, will be utterly absent.
 
21
Instead of looking to published attempts to systematically present AI (such as the textbooks upon which we rely herein), one could survey practitioners in AI, and see if their views harmonize with the publications explicitly designed to present all of AI (from a high-altitude perspective). E.g., one could turn to such reports as (Müller and Bostrom 2016), in which the authors report on a specific question, given at a conference that celebrated AI’s “turning 50” (AI@50), which asked for an opinion as to the earliest date (computing) machines would be able to simulate human-level learning. It’s rather interesting that 41% of respondents said this would never happen. It would be interesting to know if, in the context of the attention ML receives these days, the number of these pessimists would be markedly smaller. If so, that may well be because, intuitively, plenty of people harbor suspicions that ML in point of fact hasn’t achieved any human-level real learning.
 
22
Luger’s book revolves around a fundamental distinction between what he calls weak problem-solving versus strong problem-solving.
 
23
There are a few exceptions. Hummel (2010) has explained that sophisticated and powerful forms of symbolic learning, ones aligned with second-order logic, are superior to associative forms of learning. Additionally, there’s one clear historical exception, but it’s now merely a sliver in psychology (specifically, in psychology of reasoning), and hence presently has insufficient adherents to merit inclusion in the ontology we now proceed to canvass. We refer here to the type of learning over the years of human development and formal education posited by Piaget; e.g. see (Inhelder and Piaget 1958). Piaget’s view, in a barbaric nutshell, is that, given solid academic education, nutrition, and parenting, humans develop the capacity to reason with and even eventually over first-order and modal logic—which means that such humans would develop the capacity to learn in \(\mathcal {RL}\) fashion, in school. Since attacks on Piaget’s view, starting originally with those of Wason and Johnson-Laird (e.g. see Wason and Johnson-Laird 1972), many psychologists have rejected Piaget’s position. For what it’s worth, Bringsjord has defended Piaget; see e.g. (Bringsjord et al. 1998).
 
24
We are happy to concede that years of laborious (and tedious?) study of conditioning using appetitive and aversive reinforcement (and such phenomena as inhibitory conditioning, conditioned suppression, higher-order conditioning, conditioned reinforcement, and blocking) has revealed that conditioning can’t be literally reduced to new reflexes, but there is no denying that in conditioning, any new knowledge and representation that takes form falls light years short of \(\mathcal {RL}\).
 
25
Note that all occurrences of ‘understanding’ in the itemized list that follows, in keeping with the psychometric operationalization introduced at the outset in order not to rely on the murky concept of understanding, could be invoked here; but doing so would take much space and time, and be quite inelegant.
 
26
Peano Arithmetic (PA) is rarely introduced by name in K–12 education, but all the axioms of it, save perhaps for the Induction Schema, are introduced and taught there.
 
27
This conception matches that of an agent in orthodox AI: see the textbooks, e.g. (Luger 2008; Russell and Norvig 2009).
 
Literatur
Zurück zum Zitat Achab, M., Bacry, E., Gaïffas, S., Mastromatteo, I., Muzy, J.F.: Uncovering causality from multivariate Hawkes integrated cumulants. In: Precup, D., Teh, Y.W. (eds) Proceedings of the 34th International Conference on Machine Learning, PMLR, International Convention Centre, Sydney, Australia. Proceedings of Machine Learning Research, vol. 70, pp. 1–10 (2017). http://proceedings.mlr.press/v70/achab17a.html Achab, M., Bacry, E., Gaïffas, S., Mastromatteo, I., Muzy, J.F.: Uncovering causality from multivariate Hawkes integrated cumulants. In: Precup, D., Teh, Y.W. (eds) Proceedings of the 34th International Conference on Machine Learning, PMLR, International Convention Centre, Sydney, Australia. Proceedings of Machine Learning Research, vol. 70, pp. 1–10 (2017). http://​proceedings.​mlr.​press/​v70/​achab17a.​html
Zurück zum Zitat Arkoudas, K.: Denotational proof languages. Ph.D. thesis, MIT (2000) Arkoudas, K.: Denotational proof languages. Ph.D. thesis, MIT (2000)
Zurück zum Zitat Arkoudas, K., Bringsjord, S.: Computers, justification, and mathematical knowledge. Minds Mach. 17(2), 185–202 (2007)CrossRef Arkoudas, K., Bringsjord, S.: Computers, justification, and mathematical knowledge. Minds Mach. 17(2), 185–202 (2007)CrossRef
Zurück zum Zitat Arkoudas, K., Musser, D.: Fundamental Proof Methods in Computer Science: A Computer-Based Approach. MIT Press, Cambridge (2017)MATH Arkoudas, K., Musser, D.: Fundamental Proof Methods in Computer Science: A Computer-Based Approach. MIT Press, Cambridge (2017)MATH
Zurück zum Zitat Bandura, A., Walters, R.H.: Social Learning Theory, vol. 1. Prentice-Hall, Englewood Cliffs (1977) Bandura, A., Walters, R.H.: Social Learning Theory, vol. 1. Prentice-Hall, Englewood Cliffs (1977)
Zurück zum Zitat Bandura, A., Ross, D., Ross, S.A.: Transmission of aggression through imitation of aggressive models. J. Abnorm. Soc. Psychol. 63(3), 575 (1961)CrossRef Bandura, A., Ross, D., Ross, S.A.: Transmission of aggression through imitation of aggressive models. J. Abnorm. Soc. Psychol. 63(3), 575 (1961)CrossRef
Zurück zum Zitat Barrett, L.: Beyond the Brain: How Body and Environment Shape Animal and Human Minds. Princeton University Press, Princeton (2015) Barrett, L.: Beyond the Brain: How Body and Environment Shape Animal and Human Minds. Princeton University Press, Princeton (2015)
Zurück zum Zitat Bellman, A., Bragg, S., Handlin, W.: Algebra 2: Common Core. Pearson, Upper Saddle River (2012). Series Authors: Charles, R., Kennedy, D., Hall, B., Consulting Authors: Murphy, S.G Bellman, A., Bragg, S., Handlin, W.: Algebra 2: Common Core. Pearson, Upper Saddle River (2012). Series Authors: Charles, R., Kennedy, D., Hall, B., Consulting Authors: Murphy, S.G
Zurück zum Zitat Boolos, G.S., Burgess, J.P., Jeffrey, R.C.: Computability and Logic, 4th edn. Cambridge University Press, Cambridge (2003)MATH Boolos, G.S., Burgess, J.P., Jeffrey, R.C.: Computability and Logic, 4th edn. Cambridge University Press, Cambridge (2003)MATH
Zurück zum Zitat Bringsjord, S.: What Robots Can and Can’t Be. Kluwer, Dordrecht (1992)CrossRef Bringsjord, S.: What Robots Can and Can’t Be. Kluwer, Dordrecht (1992)CrossRef
Zurück zum Zitat Bringsjord, S.: Psychometric artificial intelligence. J. Exp. Theor. Artif. Intell. 23(3), 271–277 (2011)CrossRef Bringsjord, S.: Psychometric artificial intelligence. J. Exp. Theor. Artif. Intell. 23(3), 271–277 (2011)CrossRef
Zurück zum Zitat Bringsjord, S., Arkoudas, K.: The modal argument for hypercomputing minds. Theor. Comput. Sci. 317, 167–190 (2004)MathSciNetCrossRef Bringsjord, S., Arkoudas, K.: The modal argument for hypercomputing minds. Theor. Comput. Sci. 317, 167–190 (2004)MathSciNetCrossRef
Zurück zum Zitat Bringsjord, S., Noel, R.: Real robots and the missing thought experiment in the Chinese room dialectic. In: Preston, J., Bishop, M. (eds.) Views into the Chinese Room: New Essays on Searle and Artificial Intelligence, pp. 144–166. Oxford University Press, Oxford (2002) Bringsjord, S., Noel, R.: Real robots and the missing thought experiment in the Chinese room dialectic. In: Preston, J., Bishop, M. (eds.) Views into the Chinese Room: New Essays on Searle and Artificial Intelligence, pp. 144–166. Oxford University Press, Oxford (2002)
Zurück zum Zitat Bringsjord, S., Zenzen, M.: Superminds: People Harness Hypercomputation, and More. Kluwer Academic Publishers, Dordrecht (2003)CrossRef Bringsjord, S., Zenzen, M.: Superminds: People Harness Hypercomputation, and More. Kluwer Academic Publishers, Dordrecht (2003)CrossRef
Zurück zum Zitat Bringsjord, S., Bringsjord, E., Noel, R.: In defense of logical minds. In: Proceedings of the 20th Annual Conference of the Cognitive Science Society, pp. 173–178. Lawrence Erlbaum, Mahwah (1998) Bringsjord, S., Bringsjord, E., Noel, R.: In defense of logical minds. In: Proceedings of the 20th Annual Conference of the Cognitive Science Society, pp. 173–178. Lawrence Erlbaum, Mahwah (1998)
Zurück zum Zitat Bringsjord, S., Kellett, O., Shilliday, A., Taylor, J., van Heuveln, B., Yang, Y., Baumes, J., Ross, K.: A new Gödelian argument for hypercomputing minds based on the busy beaver problem. Appl. Math. Comput. 176, 516–530 (2006)MathSciNetMATH Bringsjord, S., Kellett, O., Shilliday, A., Taylor, J., van Heuveln, B., Yang, Y., Baumes, J., Ross, K.: A new Gödelian argument for hypercomputing minds based on the busy beaver problem. Appl. Math. Comput. 176, 516–530 (2006)MathSciNetMATH
Zurück zum Zitat Chisholm, R.: Theory of Knowledge. Prentice-Hall, Englewood Cliffs (1966) Chisholm, R.: Theory of Knowledge. Prentice-Hall, Englewood Cliffs (1966)
Zurück zum Zitat Davis, M., Weyuker, E.: Computability, Complexity, and Languages: Fundamentals of Theoretical Computer Science, 1st edn. Academic Press, New York (1983)MATH Davis, M., Weyuker, E.: Computability, Complexity, and Languages: Fundamentals of Theoretical Computer Science, 1st edn. Academic Press, New York (1983)MATH
Zurück zum Zitat Domjan, M.: The Principles of Learning and Behavior, 7th edn. Cengage Learning, Stamford (2015) Domjan, M.: The Principles of Learning and Behavior, 7th edn. Cengage Learning, Stamford (2015)
Zurück zum Zitat Huitt, W.: Classroom Instruction. Educational Psychology Interactive (2003) Huitt, W.: Classroom Instruction. Educational Psychology Interactive (2003)
Zurück zum Zitat Inhelder, B., Piaget, J.: The Growth of Logical Thinking from Childhood to Adolescence. Basic Books, New York (1958)CrossRef Inhelder, B., Piaget, J.: The Growth of Logical Thinking from Childhood to Adolescence. Basic Books, New York (1958)CrossRef
Zurück zum Zitat Jain, S., Osherson, D., Royer, J., Sharma, A.: Systems That Learn: An Introduction to Learning Theory, 2nd edn. MIT Press, Cambridge (1999) Jain, S., Osherson, D., Royer, J., Sharma, A.: Systems That Learn: An Introduction to Learning Theory, 2nd edn. MIT Press, Cambridge (1999)
Zurück zum Zitat Kearns, M., Vazirani, U.: An Introduction to Computational Learning Theory. MIT Press, Cambridge (1994) Kearns, M., Vazirani, U.: An Introduction to Computational Learning Theory. MIT Press, Cambridge (1994)
Zurück zum Zitat Kitzelmann, E.: Inductive programming: a survey of program synthesis techniques. In: International Workshop on Approaches and Applications of Inductive Programming, pp 50–73. Springer (2009) Kitzelmann, E.: Inductive programming: a survey of program synthesis techniques. In: International Workshop on Approaches and Applications of Inductive Programming, pp 50–73. Springer (2009)
Zurück zum Zitat LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)CrossRef LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)CrossRef
Zurück zum Zitat Luger, G.: Artificial Intelligence: Structures and Strategies for Complex Problem Solving, 6th edn. Pearson, London (2008) Luger, G.: Artificial Intelligence: Structures and Strategies for Complex Problem Solving, 6th edn. Pearson, London (2008)
Zurück zum Zitat Mackintosh, N.J.: Conditioning and Associative Learning. Calendron Press, Oxford (1983) Mackintosh, N.J.: Conditioning and Associative Learning. Calendron Press, Oxford (1983)
Zurück zum Zitat Moore, R.C.: Making the transition to formal proof. Educ. Stud. Math. 27(3), 249–266 (1994)CrossRef Moore, R.C.: Making the transition to formal proof. Educ. Stud. Math. 27(3), 249–266 (1994)CrossRef
Zurück zum Zitat Müller, V., Bostrom, N.: Future progress in artificial intelligence: a survey of expert opinion. In: Müller, V. (ed.) Fundamental Issues of Artificial Intelligence (Synthese Library), pp. 553–571. Springer, Berlin (2016) Müller, V., Bostrom, N.: Future progress in artificial intelligence: a survey of expert opinion. In: Müller, V. (ed.) Fundamental Issues of Artificial Intelligence (Synthese Library), pp. 553–571. Springer, Berlin (2016)
Zurück zum Zitat Penn, D., Holyoak, K., Povinelli, D.: Darwin’s mistake: explaining the discontinuity between human and nonhuman minds. Behav. Brain Sci. 31, 109–178 (2008) Penn, D., Holyoak, K., Povinelli, D.: Darwin’s mistake: explaining the discontinuity between human and nonhuman minds. Behav. Brain Sci. 31, 109–178 (2008)
Zurück zum Zitat Putnam, H.: Trial and error predicates and a solution to a problem of Mostowski. J. Symbolic Logic 30(1), 49–57 (1965)MathSciNetCrossRef Putnam, H.: Trial and error predicates and a solution to a problem of Mostowski. J. Symbolic Logic 30(1), 49–57 (1965)MathSciNetCrossRef
Zurück zum Zitat Ross, J.: Immaterial aspects of thought. J. Philos. 89(3), 136–150 (1992)CrossRef Ross, J.: Immaterial aspects of thought. J. Philos. 89(3), 136–150 (1992)CrossRef
Zurück zum Zitat Russell, S., Norvig, P.: Artificial Intelligence: A Modern Approach, 3rd edn. Prentice Hall, Upper Saddle River (2009)MATH Russell, S., Norvig, P.: Artificial Intelligence: A Modern Approach, 3rd edn. Prentice Hall, Upper Saddle River (2009)MATH
Zurück zum Zitat Schapiro, A., Turk-Browne, N.: Statistical learning. Brain Mapp. Encyclopedic Ref. 3, 501–506 (2015)CrossRef Schapiro, A., Turk-Browne, N.: Statistical learning. Brain Mapp. Encyclopedic Ref. 3, 501–506 (2015)CrossRef
Zurück zum Zitat Searle, J.: Minds, brains and programs. Behav. Brain Sci. 3, 417–424 (1980)CrossRef Searle, J.: Minds, brains and programs. Behav. Brain Sci. 3, 417–424 (1980)CrossRef
Zurück zum Zitat Shalev-Shwartz, S., Ben-David, S.: Understanding Machine Learning: From Theory to Algorithms. Cambridge University Press, Cambridge (2014)CrossRef Shalev-Shwartz, S., Ben-David, S.: Understanding Machine Learning: From Theory to Algorithms. Cambridge University Press, Cambridge (2014)CrossRef
Zurück zum Zitat Stewart, J.: Calculus. 8th edn. Cengage Learning, Boston (2016), We refer here to an electronic version of the print textbook. The “Student Edition” of the hard-copy textbook has an ISBN 978-1-305-27176-0 Stewart, J.: Calculus. 8th edn. Cengage Learning, Boston (2016), We refer here to an electronic version of the print textbook. The “Student Edition” of the hard-copy textbook has an ISBN 978-1-305-27176-0
Zurück zum Zitat Titley, H.K., Brunel, N., Hansel, C.: Toward a neurocentric view of learning. Neuron 95(1), 19–32 (2017)CrossRef Titley, H.K., Brunel, N., Hansel, C.: Toward a neurocentric view of learning. Neuron 95(1), 19–32 (2017)CrossRef
Zurück zum Zitat Wason, P., Johnson-Laird, P.: Psychology of Reasoning: Structure and Content. Harvard University Press, Cambridge (1972) Wason, P., Johnson-Laird, P.: Psychology of Reasoning: Structure and Content. Harvard University Press, Cambridge (1972)
Zurück zum Zitat Wolpert, D.H.: The lack of a priori distinctions between learning algorithms. Neural Comput. 8(7), 1341–1390 (1996)CrossRef Wolpert, D.H.: The lack of a priori distinctions between learning algorithms. Neural Comput. 8(7), 1341–1390 (1996)CrossRef
Metadaten
Titel
Do Machine-Learning Machines Learn?
verfasst von
Selmer Bringsjord
Naveen Sundar Govindarajulu
Shreya Banerjee
John Hummel
Copyright-Jahr
2018
DOI
https://doi.org/10.1007/978-3-319-96448-5_14