Skip to main content

Part of the book series: Philosophical Studies Series ((PSSP,volume 122))

Abstract

Is possible to understand robots as agents with moral responsibility? It will first discuss the importance and practical relevance of this question which is due to the growing autonomous agency of robots. I will then use Daniel Dennett’s theory of personhood to argue that at the moment we have no reason to see robots as responsible agents. However it might still be possible to develop downgraded versions of moral agency that fit robots better than the traditional concept of moral responsibility. Such accounts are developed by Colin Allen and Wendel Wallach, James Moor as well as Luciano Floridi and J.W. Sanders. I will argue that those attempts fail, at least if applied to the current development of robotics, because they underestimate the practical character of morality and the problem of reasonable moral disagreement. In the end I will point out that it still is possible to integrate robots into responsible groups without conceptualizing them as moral agents themselves.

Adama: “She was a Cylon, a machine. Is that what Boomer was, a machine? A thing?”

Tyrol: “That’s what she turned out to be.”

Adama: “She was more than that to us. She was more than that to me. She was a vital, living person aboard my ship for almost two years. She couldn’t have been just a machine. Could you love a machine?”

(Battlestar Galactica, ‘The Farm’, Season 2, Episode 5, 2005)

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 54.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The position of humankind in the cosmos is unquestionably one of the fundamental anthropological questions and is an essential part of the question: “What is the human?” as discusses by philosophical anthropologists like Helmut Plessner (1986) and Max Scheler (2009).

  2. 2.

    Catrin Misselhorn (2013) disagrees. She thinks that for moral agency it is enough to be able to act for moral reasons. Small children or animals acting out of sympathy would be moral agents in this sense and robots could be too. The problem with such a downgraded concept of moral agency is that it neglects reasonable disagreement regarding moral reasons, as I will argue later on. Moral agents need to be able to deal with this reasonable disagreement by deliberating about the moral value of her reasons. Otherwise they do not act for moral reasons, but only for reasons that happen to be in line with morality.

  3. 3.

    However, it is quite debatable whether animals can be said to have a form of dignity (certainly not human dignity), but perhaps an animal dignity or one shared with humans, such as the dignity of the creature, as it is stated in the constitution of Switzerland.

  4. 4.

    Floridi and Sanders (2004, 350) argue that there must be some connection between being a moral agent and being a moral patient. Contrary to this I believe that it is possible that being a moral agent and being a moral patient is connected to totally different properties, rationality and sentience for instance. Then there is no difficulty in assuming that there might be entities that are moral agents, but no moral patients and vice versa.

  5. 5.

    This is often called the merit-based concept of responsibility, because the person or persons responsible deserves a punishment or reward. See: John M. Fischer and Mark Ravizza, Responsibility and Control: A Theory of Moral Responsibility, Cambridge: Cambridge University Press 1998, pp. 207 ff.

  6. 6.

    In the Nicomachean Ethics (1110a–1111b4), Aristotle already pointed to the control condition and the epistemic condition regarding information.

  7. 7.

    It is not enough to say that language has a propositional content, which in a sense also applies to the signals emitted by animals. It is more about the intentional content of language compared to its extensional content (Frege 1962). Lewis Carroll, the famous author of Alice in Wonderland, and the famous logician Charles L. Dodgson are the same person. The two names have therefore the same extension, but do not have the same intension, because a speaker can refer to both names without knowing that the two names refer to the same person.

  8. 8.

    This is being controversially discussed with the ‘cleverbot’ programme. This programme is free to use online. If you know that it is a machine, it is not very difficult to see that it is not human. Most recently the programme name Eugene Goostman is said to have passed the test.

  9. 9.

    Of course, this is rather an empirical assertion, which goes back to something like a discursive impression.

References

  • Abney, Keith. 2012. Robotics, ethical theory, and metaethics: A guide for the perplexed. In Robot ethics. The ethical and social implications of robotics, ed. Patrick Lin, Keith Abney, and George A. Bekey, 35–52. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Allen, Colin, and Wendel Wallach. 2010. Moral machines: Teaching robots right from wrong. Oxford: Oxford University Press.

    Google Scholar 

  • Allen, Colin, and Wendel Wallach. 2012. Moral machines. Contradiction in terms or abdication of human responsibility? In Robot ethics. The ethical and social implications of robotics, ed. Patrick Lin, Keith Abney, and George A. Bekey, 55–68. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Allen, Colin, Gary Varner, and Jason Zinser. 2000. Prolegomena to any future artificial moral agent. Journal of Experimental & Theoretical Artificial Intelligence 12: 251–261.

    Article  Google Scholar 

  • Anscombe, Elisabeth. 1957. Intention. Cambridge, MA: Harvard University Press.

    Google Scholar 

  • Arkin, Ronald. 2009. Governing lethal behavior in autonomous robots. Boca Raton: Chapman and Hall.

    Book  Google Scholar 

  • Asaro, Peter M. 2012. A body to kick, but still no soul to damn: Legal perspectives on robots. In Robot ethics. The ethical and social implications of robotics, ed. Patrick Lin, Keith Abney, and George A. Bekey, 169–186. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Beck, Susanne. 2009. Grundlegende Fragen zum rechtlichen Umgang mit der Robotik. Juristische Rundschau 1(6): 225–230.

    Google Scholar 

  • Bekey, George A. 2012. Current trends in robotics: Technology and ethics. In Robot ethics. The ethical and social implications of robotics, ed. Patrick Lin, Keith Abney, and George A. Bekey, 17–34. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Benford, Gregory, and Elizabeth Malartre. 2007. Beyond human: Living with robots and cyborgs. New York: St. Martin’s Press.

    Google Scholar 

  • Birnbacher, Dieter. 1995. Tun und Unterlassen. Stuttgart: Reclam.

    Google Scholar 

  • Birnbacher, Dieter. 2006. Das Dilemma des Personenbegriffs. In Bioethik zwischen Natur und Interesse, ed. Birnbacher Dieter, 53–76. Frankfurt am Main: Suhrkamp.

    Google Scholar 

  • Crnkovic, Gordana Dodig, and Baran Çürüklü. 2012. Robots ethical by design. Ethics and Information Technology 14: 61–71.

    Article  Google Scholar 

  • Dancy, Jonathan. 2006. Ethics without principles. Oxford: Oxford University Press.

    Google Scholar 

  • Dennett, Daniel C. 1976. Conditions of personhood. In The identities of persons, ed. Rorty Amelie Oksenberg, 175–196. Berkeley: University of California Press.

    Google Scholar 

  • Dennett, Daniel. 1987. The intentional stance. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Dennett, Daniel. 1992. Consciousness explained. Boston: Back Bay Books.

    Google Scholar 

  • Dennett, Daniel. 1997. When HAL kills, who’s to blame? Computer ethics. In HAL’s legacy: 2001’s computer as dream and reality, ed. David G. Stork, 351–366. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Floridi, Luciano, and J.W. Sanders. 2004. On the morality of artificial agents. Minds and Machines 14(3): 349–379.

    Article  Google Scholar 

  • Frankfurt, Harry. 1988. The importance of what we care about. Cambridge: Cambridge University Press.

    Book  Google Scholar 

  • Frege, Gottlob. 1962. Über Sinn und Bedeutung. In Funktion, Begriff, Bedeutung. Fünf logische Studien, 38–63. Göttingen: Vandenhoeck & Ruprecht.

    Google Scholar 

  • Gunkel, David J. 2012. The machine question: Critical perspectives on AI, robots, and ethics. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Johnson, Deborah G., and Keith W. Miller. 2008. Un-making artificial moral agents. Ethics and Information Technology 10: 123–133.

    Article  Google Scholar 

  • Jonsen, Albert R., and Stephen Toulmin. 1988. The abuse of casuistry: A history of moral reasoning. Berkeley: University of California Press.

    Google Scholar 

  • Lin, Patrick. 2012. Introduction to robot ethics. In Robot ethics. The ethical and social implications of robotics, ed. Patrick Lin, Keith Abney, and George A. Bekey, 3–16. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Miller, David. 2001. Distributing responsibilities. Journal of Political Philosophy 9(4): 453–471.

    Article  Google Scholar 

  • Miller, David. 2011. Taking up the slack? Responsibility and justice in situations of partial compliance. In Responsibility and distributive justice, ed. Carl Knight and Zofia Stemplowska, 230–245. Oxford: Oxford University Press.

    Chapter  Google Scholar 

  • Misselhorn, Catrin. 2013. Robots as moral agents. In Roboethics, proceedings of the annual conference on ethics of the German association for social science research on Japan, ed. Frank Roevekamp, 30–42. Iudicum: München.

    Google Scholar 

  • Moor, James H. 2006. The nature, importance, and difficulty of machine ethics. IEEE Intelligent Systems 21(4): 18–21.

    Article  Google Scholar 

  • Moore, Michael S. 2010. Causation and responsibility: An essay in law, morals, and metaphysics. Oxford: Oxford University Press.

    Google Scholar 

  • Neuhäuser, Christian. 2011. Unternehmen als moralische Akteure. Berlin: Suhrkamp.

    Google Scholar 

  • Neuhäuser, Christian. 2012. Künstliche Intelligenz und ihr moralischer Standpunkt. In Jenseits von Mensch und Maschine, ed. Beck Susanne, 23–42. Baden Baden: Nomos.

    Google Scholar 

  • Nourbakhsh, Illah Reza. 2013. Robot futures. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Plessner, Helmut. 1986. Mit anderen Augen: Aspekte einer philosophischen Anthropologie. Stuttgart: Reclam.

    Google Scholar 

  • Quante, Michael. 2007. Person. Berlin: de Gruyter.

    Book  Google Scholar 

  • Rovane, Carol. 1995. The personal stance. Philosophical Topics 22: 397–409.

    Google Scholar 

  • Scheler, Max. 2009. The human place in the cosmos. Evanston: Northwestern University Press.

    Google Scholar 

  • Sharkey, Noel. 2009. Death strikes from the sky: The calculus of proportionality. IEEE Science and Society 2009: 16–19.

    Google Scholar 

  • Sharkey, Noel. 2012. Killing made easy: From joysticks to politics. In Robot ethics. The ethical and social implications of robotics, ed. Patrick Lin, Keith Abney, and George A. Bekey, 111–128. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Sharkey, Noel, and Sharkey Amanda. 2012. The rights and wrongs of robot. In Robot ethics. The ethical and social implications of robotics, ed. Lin Patrick, Abney Keith, and George A. Bekey, 267–282. Cambridge, MA: The MIT Press.

    Google Scholar 

  • Singer, Peter W. 2009. Wired for war: The robotics revolution and conflict in the 21st century. New York: The Penguin Press.

    Google Scholar 

  • Smith, Christian. 2010. What is a person? Rethinking humanity, social life, and the moral good from the person up. Chicago: University of Chicago Press.

    Book  Google Scholar 

  • Sullins, John P. 2006. When is a robot a moral agent? International Review of Information Ethics 6(12): 23–30.

    Google Scholar 

  • Thomson, Judith Jarvis. 1985. The trolley problem. Yale Law Journal 94: 1395–1415.

    Article  Google Scholar 

  • Watson, Gary. 1975. Free agency. Journal of Philosophy 72: 205–220.

    Article  Google Scholar 

  • Wild, Markus. 2011. Tierphilosophie zur Einführung. Hamburg: Junius Verlag.

    Google Scholar 

  • Young, Iris Marion. 2006. Responsibility and global justice: A social connection model. Social Philosophy and Policy 23: 102–130.

    Article  Google Scholar 

  • Young, Iris Marion. 2011. Responsibility for justice. New York: Oxford University Press.

    Book  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Christian Neuhäuser .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this chapter

Cite this chapter

Neuhäuser, C. (2015). Some Sceptical Remarks Regarding Robot Responsibility and a Way Forward. In: Misselhorn, C. (eds) Collective Agency and Cooperation in Natural and Artificial Systems. Philosophical Studies Series, vol 122. Springer, Cham. https://doi.org/10.1007/978-3-319-15515-9_7

Download citation

Publish with us

Policies and ethics