Skip to main content
Top
Published in: Ethics and Information Technology 1/2019

03-11-2018 | Original Paper

The value alignment problem: a geometric approach

Author: Martin Peterson

Published in: Ethics and Information Technology | Issue 1/2019

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Stuart Russell defines the value alignment problem as follows: How can we build autonomous systems with values that “are aligned with those of the human race”? In this article I outline some distinctions that are useful for understanding the value alignment problem and then propose a solution: I argue that the methods currently applied by computer scientists for embedding moral values in autonomous systems can be improved by representing moral principles as conceptual spaces, i.e. as Voronoi tessellations of morally similar choice situations located in a multidimensional geometric space. The advantage of my preferred geometric approach is that it can be implemented without specifying any utility function ex ante.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Footnotes
1
Tesla’s autopilot mode is marketed as a semi-autonomous system, not as a fully autonomous one.
 
2
For an overview of al three accidents, see The Guardian (March 31, 2018).
 
3
I leave it open whether autonoumous systems make decisions, or if all decisions are ultimately made by the engineers who design these systems. For the purposes of this paper there is no need to ascribe moral agency to autonomous systems.
 
4
Christopher von Hugo, manager of driver assistance and active safety at Mercedes-Benz, announced at the Paris auto show in 2016 that autonomous vehicles should always prioritize occupant safety over pedestrians. See Taylor (2016). It leave it to the reader to determine whether Mr. Hugo was speaking on behalf of his employer or merely expressing his personal opinion.
 
5
See e.g. Goodall (2016) Carfwod (2016), but note that Nyholm and Smids (2016) question the analogy.
 
6
See Bostrom (2014) for an extensive discussion of this topic. See also Dafoe and Russell (2016).
 
7
The quote is from a talk Dr. Russell gave at the World Economic Forum in Davos, Switzerland in Januray 2015. The talk is available on Youtube (https://​www.​youtube.​com/​watch?​v=​WvmeTaFc_​Qw). Russell has also expressed the same idea in the papers listed in the references.
 
8
Russell (2016. p. 59).
 
9
See e.g. Bostrom (2014) and Milli et al. (2017).
 
10
If an ethical theory ranks some options as infinitely better than others, or entails cyclical orderings, then no real-valued utility function could mimic the prescriptions of such an ethical theory. It is also an open question whether the “theory” I sketch in this article could be represented by some real-valued utility function. (This depends on how we understand the ranking of domain-specific principles.) Brown (2011) also points out that no real-valued utility function can account for the existence of moral dilemmas. See Peterson (2013, Chap. 8) for a discussion of how hyper-real utility functions could help us overcome this problem.
 
11
For reasons explained in the previous footnote, a problem with this suggestion might be that no real-valued utility function can account for Aristotle’s notion of supererogation. See Peterson (2013, Chap. 8).
 
12
IEEE (2017a).
 
13
IEEE (2017a, pp. 23, 36).
 
14
IEEE (2017a, p. 20).
 
15
For an overview, see Attfiled (2014).
 
16
Hadfield-Menell et al. (2016, p. 2).
 
17
Milli et al. (2017, p. 1).
 
18
IEEE (2017b, p. 1).
 
19
This is a fundamenatal assumption in Bostrom (2014) and, for instance, Milli et al. (2017), but it has far as I am aware never been exstensively discussed.
 
20
For reasons explained in the previous footnote, a problem with this suggestion might be that no real-valued utility function can account for Aristotle’s notion of supererogation. See Peterson (2013, Chap. 8).
 
21
Whether my proposal can be mimicked by some real-valued utility function is an open question (as noted in footnote 10), and also irrelevant. What matters is that my proposal can be implemented in a machine without explicitly ascribing utilities to outcomes or alternatives. From an epistemic point of this, this is a clear advantage over the utility-based approach.
 
22
The section draws on Chapter 1 in ET.
 
23
See ET, pp. 14–15.
 
24
See Nicomachean Ethics 1131a10–b15; Politics, III.9.1280 a8–15, III. 12. 1282b18–23.
 
25
See Jonsen and Toulmin (1988) for a defense of causuistry.
 
26
CBA: An option is morally right only if the net surplus of benefits over costs for all those affected is at least as large as that of every alternative.
 
27
PP: An option is morally right only if reasonable precautionary measures are taken to safeguard against uncertain but non-negligible threats.
 
28
ST: An option is morally right only if it does not lead to any significant long-term depletion of natural, social or economic resources.
 
29
AUT: An option is morally right only if it does not reduce the independence, self-governance or freedom of the people affected by it.
 
30
FP: An option is morally right only if it does not lead to unfair inequalities among the people affected by it.
 
31
Note that I am not claiming that all ethical theories are false. I am merely suggesting that it is not neceesary to take a stance on which theory is correct in order to align the values of autonous systems with ours in the manner specified in the moderate value alignment thesis.
 
32
It is of course possible that the majority is wrong. I am not trying to derive an “ought” from an is’; see Chap. 3 of ET for a discussion of Hume’s Is-Ought principle.
 
33
A reviwer has suggested that it would be helpful to clarify how the geometric method differs from Rawls’ method of reflective equilibrium. The most important difference is that unlike Rawls’ method, the geomtric method is compatible with coherentistic as well as foundationalist principles. The ex ante mechanism for selecting paradigm cases outlined in Chapter 2 of ET assigns a priviliged, foundational role to paradigm cases. The ex post mechanism discussed in the same chapter is coherentistic in the sense that the location of the paradigm cases depends on what cases the principle has been applied to in the past.
 
34
See Chapter 8 of ET.
 
35
See Chapters 1 and 2. See also the experimental evidence report in Chapters 3 and 5.
 
36
See e.g. Peterson (2013) for a defense of this view.
 
37
I would like to thank Rob Reed for suggseting this helpful point to me.
 
38
See, for instance, Gavagai.se.
 
39
Shrader-Frechette (2017).
 
40
Peterson (2017, pp. 37–38).
 
41
Stewart et al. (1973, pp. 415–417), my italics.
 
42
Kruskal and Wish (1978, pp. 30–31), my italics.
 
43
Lokhorst (2018, p. 1).
 
44
ET, p. 17.
 
45
Peterson (2017, p. 17).
 
Literature
go back to reference Anderson, M., & Anderson, S. L. (2014). GenEth: A general ethical dilemma analyzer.” Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence (2014): 253–261. Anderson, M., & Anderson, S. L. (2014). GenEth: A general ethical dilemma analyzer.” Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence (2014): 253–261.
go back to reference Attfield, R. (2014). Environmental ethics: An overview for the twenty-first century. New York: Wiley. Attfield, R. (2014). Environmental ethics: An overview for the twenty-first century. New York: Wiley.
go back to reference Bostrom, N. (2014). Superintelligence. Oxford: Oxford University Press. Bostrom, N. (2014). Superintelligence. Oxford: Oxford University Press.
go back to reference Crawford, K., & Calo, R. (2016). There is a blind spot in AI research. Nature, 538(7625). Crawford, K., & Calo, R. (2016). There is a blind spot in AI research. Nature, 538(7625).
go back to reference Dafoe, A., & Russell, S. (2016). Yes, we are worried about the existential risk of artificial intelligence. MIT Technology Review. Dafoe, A., & Russell, S. (2016). Yes, we are worried about the existential risk of artificial intelligence. MIT Technology Review.
go back to reference Gärdenfors, P. (2000). Conceptual spaces: The geometry of thought. Cambridge: MIT Press.CrossRef Gärdenfors, P. (2000). Conceptual spaces: The geometry of thought. Cambridge: MIT Press.CrossRef
go back to reference Gärdenfors, P. (2014). The geometry of meaning: Semantics based on conceptual spaces. Cambridge: MIT Press.MATH Gärdenfors, P. (2014). The geometry of meaning: Semantics based on conceptual spaces. Cambridge: MIT Press.MATH
go back to reference Goodall, N. J. (2016). Can you program ethics into a self-driving car? IEEE Spectrum, 53(6), 28–58.CrossRef Goodall, N. J. (2016). Can you program ethics into a self-driving car? IEEE Spectrum, 53(6), 28–58.CrossRef
go back to reference Guardian Staff and Agencies, (2018). Tesla car that crashed and killed driver was running on Autopilot, firm says. The Guardian, March 31st, 2018. Guardian Staff and Agencies, (2018). Tesla car that crashed and killed driver was running on Autopilot, firm says. The Guardian, March 31st, 2018.
go back to reference Hadfield-Menell, D., Dragan, A., Abbeel, P., & Russell, S. (2016). “The off-switch game”, arXiv preprint arXiv: 1611.08219. Hadfield-Menell, D., Dragan, A., Abbeel, P., & Russell, S. (2016). “The off-switch game”, arXiv preprint arXiv: 1611.08219.
go back to reference Jonsen, A. R., & Toulmin, S. E. (1988). The abuse of casuistry: A history of moral reasoning. University of California Press. Jonsen, A. R., & Toulmin, S. E. (1988). The abuse of casuistry: A history of moral reasoning. University of California Press.
go back to reference Kruskal, J. B., & Wish, M. (1978). Multidimensional scaling. New York: Sage Publications.CrossRef Kruskal, J. B., & Wish, M. (1978). Multidimensional scaling. New York: Sage Publications.CrossRef
go back to reference Milli, S., Hadfield-Menell, D., Dragan, A., & Russell, S. (2017). “Should Robots be Obedient?”. arXiv preprint arXiv.1705.09990. Milli, S., Hadfield-Menell, D., Dragan, A., & Russell, S. (2017). “Should Robots be Obedient?”. arXiv preprint arXiv.1705.09990.
go back to reference Nyholm, S., & Smids, J. (2016). The ethics of accident-algorithms for self-driving cars: An applied trolley problem? Ethical Theory and Moral Practice, 19(5), 1275–1289.CrossRef Nyholm, S., & Smids, J. (2016). The ethics of accident-algorithms for self-driving cars: An applied trolley problem? Ethical Theory and Moral Practice, 19(5), 1275–1289.CrossRef
go back to reference Paulo, N. (2015). Casuistry as common law morality. Theoretical Medicine and Bioethics, 36(6), 373–389.CrossRef Paulo, N. (2015). Casuistry as common law morality. Theoretical Medicine and Bioethics, 36(6), 373–389.CrossRef
go back to reference Peterson, M. (2013). The dimensions of consequentialism: Ethics, equality and risk. Cambridge University Press. Peterson, M. (2013). The dimensions of consequentialism: Ethics, equality and risk. Cambridge University Press.
go back to reference Peterson, M. (2017). The ethics of technology: A geometric analysis of five moral principles. Oxford: Oxford University Press.CrossRef Peterson, M. (2017). The ethics of technology: A geometric analysis of five moral principles. Oxford: Oxford University Press.CrossRef
go back to reference Rosch, E. (1975). Cognitive reference points. Cognitive Psychology, 7, 532–547.CrossRef Rosch, E. (1975). Cognitive reference points. Cognitive Psychology, 7, 532–547.CrossRef
go back to reference Rosch, E. H. (1973). Natural categories. Cognitive Psychology, 4, 328–350.CrossRef Rosch, E. H. (1973). Natural categories. Cognitive Psychology, 4, 328–350.CrossRef
go back to reference Russell, S. (2016). Should we fear supersmart robots. Scientific American, 314(6), 58–59.CrossRef Russell, S. (2016). Should we fear supersmart robots. Scientific American, 314(6), 58–59.CrossRef
go back to reference Stewart, A., Prandy, K., & Blackburn, R. M. (1973) Measuring the class structure. Nature, 245, 415.CrossRef Stewart, A., Prandy, K., & Blackburn, R. M. (1973) Measuring the class structure. Nature, 245, 415.CrossRef
Metadata
Title
The value alignment problem: a geometric approach
Author
Martin Peterson
Publication date
03-11-2018
Publisher
Springer Netherlands
Published in
Ethics and Information Technology / Issue 1/2019
Print ISSN: 1388-1957
Electronic ISSN: 1572-8439
DOI
https://doi.org/10.1007/s10676-018-9486-0

Other articles of this Issue 1/2019

Ethics and Information Technology 1/2019 Go to the issue

Premium Partner