Skip to main content

2019 | OriginalPaper | Buchkapitel

14. Machine Learning and Irresponsible Inference: Morally Assessing the Training Data for Image Recognition Systems

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Just as humans can draw conclusions responsibly or irresponsibly, so too can computers. Machine learning systems that have been trained on data sets that include irresponsible judgments are likely to yield irresponsible predictions as outputs. In this paper I focus on a particular kind of inference a computer system might make: identification of the intentions with which a person acted on the basis of photographic evidence. Such inferences are liable to be morally objectionable, because of a way in which they are presumptuous. After elaborating this moral concern, I explore the possibility that carefully procuring the training data for image recognition systems could ensure that the systems avoid the problem. The lesson of this paper extends beyond just the particular case of image recognition systems and the challenge of responsibly identifying a person’s intentions. Reflection on this particular case demonstrates the importance (as well as the difficulty) of evaluating machine learning systems and their training data from the standpoint of moral considerations that are not encompassed by ordinary assessments of predictive accuracy.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
A few clarifications about the relationship between stereotypes and presumptuous judgment may be helpful. First, not all cases of presumptuous judgment involve stereotypes. Stereotypes involve associating an individual with a group (Blum 2004; Beeghly 2015). But it is possible to make a presumptuous judgment without relying on a group association. For instance, I might make a presumptuous judgment about a person’s intentions just on the basis of the assumption that her goals are the same as my own. Second, not all uses of stereotypes involve presumptuous judgments. This is simply because not all stereotypes are about persons’ intentions. Finally, regarding the moral features of stereotypes and presumptuous judgments: Presumptuousness, all else equal, tends to be morally undesirable, but it’s controversial whether this is true of all stereotypes. Beeghly (2015) argues that not all stereotyping is morally objectionable, and Lippmann (1922) saw positive and negative aspects of stereotyping. In contrast, Blum (2004) holds that stereotyping is always morally objectionable to some degree. My contention here, that presumptuous judgments manifest inadequate respect for persons as individuals, is consistent with Beeghly’s explanation of when and how stereotypes fail to respect persons as individuals. However, my thinking about why such a failure of respect is morally objectionable shares more with Blum’s analysis than with Beeghly’s. In the context of the present paper—with its focus on the moral evaluation of training data for machine learning systems—it is enough for my purposes if at least some judgments are morally objectionable precisely because of their presumptuousness.
 
2
As Hodosh et al. (2013) point out, “Gricean maxims of relevance and quantity entail that image captions that are written for people usually provide precisely the kind of information that could not be obtained from the image itself, and thus tend to bear only a tenuous relation to what is actually depicted.”
 
3
Though crowdwork raises ethical issues of its own (Marvit 2014).
 
4
This comes from the online appendix to Hodosh et al. (2013).
 
5
This suggests another way to explain what is wrong with presumptuous judgment. To judge a person’s mental states according to a standard like we would use for any other sort of judgment not involving persons, is to take what Peter Strawson (1962) called the “objective attitude” rather than the “participant attitude” toward the person.
 
6
Of course, the image recognition system could report the falling water, and we could rely on some other process to infer from the falling water that it must be raining. But this would be to limit too much the capacities of image recognition systems. A scene can be one that looks rainy, and looking rainy may be both more intuitive and more useful information than the report that it looks like water is falling from above.
 
7
There’s nothing special about the specific probability values of 0.02 and 0.98, besides the former being small and the latter being large. These values are just convenient for purposes of illustration. Values of 0.01 and 0.99 or 0.05 and 0.95 would have worked just as well (although values that were too extreme or too moderate would indeed alter the examples).
 
8
I do not intend this as a criticism of the Flickr 8k data set. Violations of the instruction I am recommending seem to appear only rarely in the data set. However, this image and the next are valuable for illustrating the worry I that is my focus.
 
9
I do not mean to imply that Dennett himself is guilty of making this assumption.
 
10
Cf. Blum (2004).
 
11
And, of course, a further worry about this strategy concerns the thorny issue about how we might go about categorizing intentions as attractive or unattractive in the first place.
 
12
Such work is already underway. See, e.g., Park et al. (unpublished ms).
 
13
Along these lines, Dennett argues, “the class of indistinguishably satisfactory models of the formal system embodied in [the] internal states [of an entity toward which we might take the intentional stance] gets smaller and smaller as we add such complexities [such as a wider range of behaviors]; the more we add, the richer or more demanding or specific the semantics of the system, until eventually we reach systems for which a unique semantic interpretation is practically (but never in principle) dictated” (1989b). Notoriously, according to both Quine and Davidson, some indeterminacy may be ineliminable. However, along with Dennett, I doubt that any remaining indeterminacy poses any practical or ethical problems in the context of machine learning systems. For discussion of indeterminacy and its (in)significance, see Davidson (1984b).
 
14
This is a specific version of the type of problem James Moor (1985) has famously called “invisibility.”
 
Literatur
Zurück zum Zitat Beeghly, Erin. 2015. What is a stereotype? What is stereotyping? Hypatia 30 (4): 675–691.CrossRef Beeghly, Erin. 2015. What is a stereotype? What is stereotyping? Hypatia 30 (4): 675–691.CrossRef
Zurück zum Zitat Blum, Lawrence. 2004. Stereotypes and stereotyping: A moral analysis. Philosophical Papers 33 (3): 251–289.MathSciNetCrossRef Blum, Lawrence. 2004. Stereotypes and stereotyping: A moral analysis. Philosophical Papers 33 (3): 251–289.MathSciNetCrossRef
Zurück zum Zitat Davidson, Donald. 1984a. Inquiries into truth and interpretation. Oxford: Clarendon Press. Davidson, Donald. 1984a. Inquiries into truth and interpretation. Oxford: Clarendon Press.
Zurück zum Zitat ———. 1984b. Belief and the basis of meaning. Reprinted in Davidson (1984a): 141–154. ———. 1984b. Belief and the basis of meaning. Reprinted in Davidson (1984a): 141–154.
Zurück zum Zitat ———. 2004b. Expressing evaluations. Reprinted in Davidson (2004a): 19–37. ———. 2004b. Expressing evaluations. Reprinted in Davidson (2004a): 19–37.
Zurück zum Zitat Dennett, Daniel. 1989a. The intentional stance. Cambridge, MA: MIT Press. Dennett, Daniel. 1989a. The intentional stance. Cambridge, MA: MIT Press.
Zurück zum Zitat ———. 1989b. True believers. Reprinted in Dennet (1989a): 13–35. ———. 1989b. True believers. Reprinted in Dennet (1989a): 13–35.
Zurück zum Zitat Fei-Fei, Li, and Li-Jia Li. 2010. What, where and who? telling the story of an image by activity classification, scene recognition and object categorization. In Computer vision, ed. Cipolla et al., 157–171. Berlin: Springer.CrossRef Fei-Fei, Li, and Li-Jia Li. 2010. What, where and who? telling the story of an image by activity classification, scene recognition and object categorization. In Computer vision, ed. Cipolla et al., 157–171. Berlin: Springer.CrossRef
Zurück zum Zitat Hodosh, Micah, Peter Young, and Julia Hockenmaier. 2013. Framing image description as a ranking task: Data, models and evaluation metrics. Journal of Artificial Intelligence Research 47: 853–899.MathSciNetCrossRef Hodosh, Micah, Peter Young, and Julia Hockenmaier. 2013. Framing image description as a ranking task: Data, models and evaluation metrics. Journal of Artificial Intelligence Research 47: 853–899.MathSciNetCrossRef
Zurück zum Zitat Karpathy, Andrej, and Li Fei-Fei. 2014. Deep visual-semantic alignments for generating image descriptions. arXiv preprint arXiv:1412.2306. Karpathy, Andrej, and Li Fei-Fei. 2014. Deep visual-semantic alignments for generating image descriptions. arXiv preprint arXiv:1412.2306.
Zurück zum Zitat Lippmann, Walter. 1922. Public opinion. New York: Macmillan. Lippmann, Walter. 1922. Public opinion. New York: Macmillan.
Zurück zum Zitat Moor, James. 1985. What is computer ethics? Metaphilosophy 16 (4): 266–275.CrossRef Moor, James. 1985. What is computer ethics? Metaphilosophy 16 (4): 266–275.CrossRef
Zurück zum Zitat Quine, W.V. 1960. Word and object. Cambridge, MA: MIT press.MATH Quine, W.V. 1960. Word and object. Cambridge, MA: MIT press.MATH
Zurück zum Zitat Rashtchian, Cyrus, Peter Young, Micah Hodosh, and Julia Hockenmaier. 2010. Collecting image annotations using Amazon’s Mechanical Turk. In Proceedings of the NAACL HLT 2010 Workshop on Creating Speech and Language Data with Amazon’s Mechanical Turk, 139–147. Association for Computational Linguistics. Rashtchian, Cyrus, Peter Young, Micah Hodosh, and Julia Hockenmaier. 2010. Collecting image annotations using Amazon’s Mechanical Turk. In Proceedings of the NAACL HLT 2010 Workshop on Creating Speech and Language Data with Amazon’s Mechanical Turk, 139–147. Association for Computational Linguistics.
Zurück zum Zitat Strawson, Peter. 1962. Freedom and resentment. Proceedings of the British Academy 48: 1–25.CrossRef Strawson, Peter. 1962. Freedom and resentment. Proceedings of the British Academy 48: 1–25.CrossRef
Zurück zum Zitat Vinyals, Oriol, Alexander Toshev, Samy Bengio, and Dumitru Erhan. 2014. Show and tell: A neural image caption generator. arXiv preprint arXiv:1411.4555. Vinyals, Oriol, Alexander Toshev, Samy Bengio, and Dumitru Erhan. 2014. Show and tell: A neural image caption generator. arXiv preprint arXiv:1411.4555.
Metadaten
Titel
Machine Learning and Irresponsible Inference: Morally Assessing the Training Data for Image Recognition Systems
verfasst von
Owen C. King
Copyright-Jahr
2019
DOI
https://doi.org/10.1007/978-3-030-01800-9_14