Skip to main content

2019 | OriginalPaper | Buchkapitel

Robot Authority in Human-Machine Teams: Effects of Human-Like Appearance on Compliance

verfasst von : Kerstin S. Haring, Ariana Mosley, Sarah Pruznick, Julie Fleming, Kelly Satterfield, Ewart J. de Visser, Chad C. Tossell, Gregory Funke

Erschienen in: Virtual, Augmented and Mixed Reality. Applications and Case Studies

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Current technology allows for the deployment of security patrol and police robots. It is expected that in the near future robots and similar technologies will exhibit some degree of authority over people within human-machine teams. Studies in classical psychology investigating compliance have shown that people tend to comply with requests from others who display or are assumed to have authority. In this study, we investigated the effect of a robot’s human-like appearance on compliance with a request. We compared two different robots to a human control condition. The robots assumed the role of a coach in learning a difficult task. We hypothesized that participants would have higher compliance with robots high compared to robots low in human-like appearance. The coach continuously prompts the participant to continue to practice the task beyond the time the participant wishes to actually proceed. Compliance was measured by time practiced after the first prompt and the total number of images processed. Results showed that compliance with the request was the highest with a human and compliance with both robots was significantly lower. However, we showed that robots can be used as persuasive coaches that can help a human teammate to persist in training task. There were no differences between the High and Low Human-Like robot for compliance time, however the Low Human-Like robot has people practise on more images than the High Human-Like robot. The implication of this study is that robots are currently inferior to humans when it comes to compliance in a human-machine team. Future robots need to be carefully designed in an authoritative way if maximizing compliance to their requests is the primary goal.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Salem, M., Dautenhahn, K.: Evaluating trust and safety in HRI: practical issues and ethical challenges. In: Emerging Policy and Ethics of Human-Robot Interaction (2015) Salem, M., Dautenhahn, K.: Evaluating trust and safety in HRI: practical issues and ethical challenges. In: Emerging Policy and Ethics of Human-Robot Interaction (2015)
2.
Zurück zum Zitat Sharkey, N.: The robot arm of the law grows longer. Computer 42(8), 115–116 (2009)CrossRef Sharkey, N.: The robot arm of the law grows longer. Computer 42(8), 115–116 (2009)CrossRef
4.
Zurück zum Zitat Dennis, L.A., Fisher, M., Lincoln, N.K., Lisitsa, A., Veres, S.M.: Practical verification of decision-making in agent-based autonomous systems. Autom. Softw. Eng. 23(3), 305–359 (2016)CrossRef Dennis, L.A., Fisher, M., Lincoln, N.K., Lisitsa, A., Veres, S.M.: Practical verification of decision-making in agent-based autonomous systems. Autom. Softw. Eng. 23(3), 305–359 (2016)CrossRef
5.
Zurück zum Zitat Li, L., Ota, K., Dong, M.: Humanlike driving: empirical decision-making system for autonomous vehicles. IEEE Trans. Veh. Technol. 67(8), 6814–6823 (2018)CrossRef Li, L., Ota, K., Dong, M.: Humanlike driving: empirical decision-making system for autonomous vehicles. IEEE Trans. Veh. Technol. 67(8), 6814–6823 (2018)CrossRef
6.
Zurück zum Zitat Cunningham, A.G., Galceran, E., Mehta, D., Ferrer, G., Eustice, R.M., Olson, E.: MPDM: multi-policy decision-making from autonomous driving to social robot navigation. In: Waschl, H., Kolmanovsky, I., Willems, F. (eds.) Control Strategies for Advanced Driver Assistance Systems and Autonomous Driving Functions. LNCIS, vol. 476, pp. 201–223. Springer, Cham (2019). https://doi.org/10.1007/978-3-319-91569-2_10CrossRef Cunningham, A.G., Galceran, E., Mehta, D., Ferrer, G., Eustice, R.M., Olson, E.: MPDM: multi-policy decision-making from autonomous driving to social robot navigation. In: Waschl, H., Kolmanovsky, I., Willems, F. (eds.) Control Strategies for Advanced Driver Assistance Systems and Autonomous Driving Functions. LNCIS, vol. 476, pp. 201–223. Springer, Cham (2019). https://​doi.​org/​10.​1007/​978-3-319-91569-2_​10CrossRef
7.
Zurück zum Zitat Long, S.K., Karpinsky, N.D., Bliss, J.P.: Trust of simulated robotic peacekeepers among resident and expatriate Americans. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 61, no. 1, pp. 2091–2095. SAGE Publications, Los Angeles (2017)CrossRef Long, S.K., Karpinsky, N.D., Bliss, J.P.: Trust of simulated robotic peacekeepers among resident and expatriate Americans. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 61, no. 1, pp. 2091–2095. SAGE Publications, Los Angeles (2017)CrossRef
8.
Zurück zum Zitat Agrawal, S., Williams, M.-A.: Robot authority and human obedience: a study of human behaviour using a robot security guard. In: Proceedings of the Companion of the 2017 ACM/IEEE International Conference on Human-Robot Interaction, pp. 57–58. ACM (2017) Agrawal, S., Williams, M.-A.: Robot authority and human obedience: a study of human behaviour using a robot security guard. In: Proceedings of the Companion of the 2017 ACM/IEEE International Conference on Human-Robot Interaction, pp. 57–58. ACM (2017)
9.
Zurück zum Zitat Hoffman, G., et al.: Robot presence and human honesty: experimental evidence. In: Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction, pp. 181–188. ACM (2015) Hoffman, G., et al.: Robot presence and human honesty: experimental evidence. In: Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction, pp. 181–188. ACM (2015)
10.
Zurück zum Zitat Benitez, J., Wyman, A.B., Carpinella, C.M., Stroessner, S.J.: The authority of appearance: how robot features influence trait inferences and evaluative responses. In: 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), pp. 397–404. IEEE (2017) Benitez, J., Wyman, A.B., Carpinella, C.M., Stroessner, S.J.: The authority of appearance: how robot features influence trait inferences and evaluative responses. In: 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), pp. 397–404. IEEE (2017)
11.
Zurück zum Zitat Inbar, O., Meyer, J.: Manners matter: trust in robotic peacekeepers. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 59, no. 1, pp. 185–189. SAGE Publications, Los Angeles (2015)CrossRef Inbar, O., Meyer, J.: Manners matter: trust in robotic peacekeepers. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 59, no. 1, pp. 185–189. SAGE Publications, Los Angeles (2015)CrossRef
12.
Zurück zum Zitat Geiskkovitch, D.Y., Cormier, D., Seo, S.H., Young, J.E.: Please continue, we need more data: an exploration of obedience to robots. J. Hum.-Robot Interact. 5(1), 82–99 (2016)CrossRef Geiskkovitch, D.Y., Cormier, D., Seo, S.H., Young, J.E.: Please continue, we need more data: an exploration of obedience to robots. J. Hum.-Robot Interact. 5(1), 82–99 (2016)CrossRef
13.
Zurück zum Zitat Milgram, S.: Behavioral study of obedience. J. Abnormal Soc. Psychol. 67(4), 371 (1963)CrossRef Milgram, S.: Behavioral study of obedience. J. Abnormal Soc. Psychol. 67(4), 371 (1963)CrossRef
14.
Zurück zum Zitat Milgram, S., Gudehus, C.: Obedience to authority (1978) Milgram, S., Gudehus, C.: Obedience to authority (1978)
15.
Zurück zum Zitat Meeus, W.H., Raaijmakers, Q.A.: Obedience in modern society: the Utrecht studies. J. Soc. Issues 51(3), 155–175 (1995)CrossRef Meeus, W.H., Raaijmakers, Q.A.: Obedience in modern society: the Utrecht studies. J. Soc. Issues 51(3), 155–175 (1995)CrossRef
16.
Zurück zum Zitat Haney, C., Banks, W.C., Zimbardo, P.G.: A study of prisoners and guards in a simulated prison. Naval Res. Rev. 9, 1–17 (1973) Haney, C., Banks, W.C., Zimbardo, P.G.: A study of prisoners and guards in a simulated prison. Naval Res. Rev. 9, 1–17 (1973)
18.
Zurück zum Zitat Weiss, D.J.: Deception by researchers is necessary and not necessarily evil. Behav. Brain Sci. 24(3), 431–432 (2001)CrossRef Weiss, D.J.: Deception by researchers is necessary and not necessarily evil. Behav. Brain Sci. 24(3), 431–432 (2001)CrossRef
19.
Zurück zum Zitat Masters, K.S.: Milgram, stress research, and the institutional review board (2009)CrossRef Masters, K.S.: Milgram, stress research, and the institutional review board (2009)CrossRef
20.
Zurück zum Zitat Haring, K.S., Watanabe, K., Velonaki, M., Tossell, C.C., Finomore, V.: FFAB-the form function attribution bias in human-robot interaction. IEEE Trans. Cogn. Dev. Syst. 10(4), 843–851 (2018)CrossRef Haring, K.S., Watanabe, K., Velonaki, M., Tossell, C.C., Finomore, V.: FFAB-the form function attribution bias in human-robot interaction. IEEE Trans. Cogn. Dev. Syst. 10(4), 843–851 (2018)CrossRef
21.
Zurück zum Zitat Haring, K.S., Matsumoto, Y., Watanabe, K.: How do people perceive and trust a lifelike robot. In: Proceedings of the World Congress on Engineering and Computer Science, vol. 1 (2013) Haring, K.S., Matsumoto, Y., Watanabe, K.: How do people perceive and trust a lifelike robot. In: Proceedings of the World Congress on Engineering and Computer Science, vol. 1 (2013)
22.
Zurück zum Zitat Haring, K.S., Watanabe, K., Silvera-Tawil, D., Velonaki, M., Takahashi, T.: Changes in perception of a small humanoid robot. In: 2015 6th International Conference on Automation, Robotics and Applications (ICARA), pp. 83–89. IEEE (2015) Haring, K.S., Watanabe, K., Silvera-Tawil, D., Velonaki, M., Takahashi, T.: Changes in perception of a small humanoid robot. In: 2015 6th International Conference on Automation, Robotics and Applications (ICARA), pp. 83–89. IEEE (2015)
23.
Zurück zum Zitat Hinds, P.J., Roberts, T.L., Jones, H.: Whose job is it anyway? A study of human-robot interaction in a collaborative task. Hum.-Comput. Interact. 19(1), 151–181 (2004)CrossRef Hinds, P.J., Roberts, T.L., Jones, H.: Whose job is it anyway? A study of human-robot interaction in a collaborative task. Hum.-Comput. Interact. 19(1), 151–181 (2004)CrossRef
24.
Zurück zum Zitat Waytz, A., Morewedge, C.K., Epley, N., Monteleone, G., Gao, J.-H., Cacioppo, J.T.: Making sense by making sentient: effectance motivation increases anthropomorphism. J. Pers. Soc. Psychol. 99(3), 410 (2010)CrossRef Waytz, A., Morewedge, C.K., Epley, N., Monteleone, G., Gao, J.-H., Cacioppo, J.T.: Making sense by making sentient: effectance motivation increases anthropomorphism. J. Pers. Soc. Psychol. 99(3), 410 (2010)CrossRef
25.
Zurück zum Zitat Waytz, A., Gray, K., Epley, N., Wegner, D.M.: Causes and consequences of mind perception. Trends Cogn. Sci. 14(8), 383–388 (2010)CrossRef Waytz, A., Gray, K., Epley, N., Wegner, D.M.: Causes and consequences of mind perception. Trends Cogn. Sci. 14(8), 383–388 (2010)CrossRef
27.
Zurück zum Zitat Phillips, E., Zhao, X., Ullman, D., Malle, B.F.: What is human-like?: Decomposing robots’ human-like appearance using the anthropomorphic robot (abot) database. In: Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction, pp. 105–113. ACM (2018) Phillips, E., Zhao, X., Ullman, D., Malle, B.F.: What is human-like?: Decomposing robots’ human-like appearance using the anthropomorphic robot (abot) database. In: Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction, pp. 105–113. ACM (2018)
28.
Zurück zum Zitat Parasuraman, R., Sheridan, T.B., Wickens, C.D.: A model for types and levels of human interaction with automation. IEEE Trans. Syst. Man Cybern.-Part A: Syst. Hum. 30(3), 286–297 (2000)CrossRef Parasuraman, R., Sheridan, T.B., Wickens, C.D.: A model for types and levels of human interaction with automation. IEEE Trans. Syst. Man Cybern.-Part A: Syst. Hum. 30(3), 286–297 (2000)CrossRef
29.
Zurück zum Zitat Beer, J.M., Fisk, A.D., Rogers, W.A.: Toward a framework for levels of robot autonomy in human-robot interaction. J. Hum.-Robot Interact. 3(2), 74–99 (2014)CrossRef Beer, J.M., Fisk, A.D., Rogers, W.A.: Toward a framework for levels of robot autonomy in human-robot interaction. J. Hum.-Robot Interact. 3(2), 74–99 (2014)CrossRef
30.
Zurück zum Zitat Onnasch, L., Wickens, C.D., Li, H., Manzey, D.: Human performance consequences of stages and levels of automation: an integrated meta-analysis. Hum. Factors 56(3), 476–488 (2014)CrossRef Onnasch, L., Wickens, C.D., Li, H., Manzey, D.: Human performance consequences of stages and levels of automation: an integrated meta-analysis. Hum. Factors 56(3), 476–488 (2014)CrossRef
31.
Zurück zum Zitat Fessler, D.M., Holbrook, C., Snyder, J.K.: Weapons make the man (larger): formidability is represented as size and strength in humans. PloS One 7(4), e32751 (2012)CrossRef Fessler, D.M., Holbrook, C., Snyder, J.K.: Weapons make the man (larger): formidability is represented as size and strength in humans. PloS One 7(4), e32751 (2012)CrossRef
32.
Zurück zum Zitat Satterfield, K., et al.: Investigating compliance in human-robot teaming. In: 2nd International Conference on Intelligent Human Systems Integration (IHSI 2019): Integrating People and Intelligent Systems (2019) Satterfield, K., et al.: Investigating compliance in human-robot teaming. In: 2nd International Conference on Intelligent Human Systems Integration (IHSI 2019): Integrating People and Intelligent Systems (2019)
33.
Zurück zum Zitat McCambridge, J., Witton, J., Elbourne, D.R.: Systematic review of the Hawthorne effect: new concepts are needed to study research participation effects. J. Clin. Epidemiol. 67(3), 267–277 (2014)CrossRef McCambridge, J., Witton, J., Elbourne, D.R.: Systematic review of the Hawthorne effect: new concepts are needed to study research participation effects. J. Clin. Epidemiol. 67(3), 267–277 (2014)CrossRef
34.
Zurück zum Zitat Wickström, G., Bendix, T.: The Hawthorne effect-what did the original Hawthorne studies actually show? Scand. J. Work Environ. Health 26, 363–367 (2000) Wickström, G., Bendix, T.: The Hawthorne effect-what did the original Hawthorne studies actually show? Scand. J. Work Environ. Health 26, 363–367 (2000)
35.
Zurück zum Zitat Parsons, H.M.: What happened at Hawthorne?: New evidence suggests the Hawthorne effect resulted from operant reinforcement contingencies. Science 183(4128), 922–932 (1974)CrossRef Parsons, H.M.: What happened at Hawthorne?: New evidence suggests the Hawthorne effect resulted from operant reinforcement contingencies. Science 183(4128), 922–932 (1974)CrossRef
36.
Zurück zum Zitat Lucas, G.M., Gratch, J., King, A., Morency, L.-P.: It’s only a computer: virtual humans increase willingness to disclose. Comput. Hum. Behav. 37, 94–100 (2014)CrossRef Lucas, G.M., Gratch, J., King, A., Morency, L.-P.: It’s only a computer: virtual humans increase willingness to disclose. Comput. Hum. Behav. 37, 94–100 (2014)CrossRef
37.
Zurück zum Zitat Linder, S.P., Nestrick, B.E., Mulders, S., Lavelle, C.L.: Facilitating active learning with inexpensive mobile robots. J. Comput. Sci. Coll. 16(4), 21–33 (2001) Linder, S.P., Nestrick, B.E., Mulders, S., Lavelle, C.L.: Facilitating active learning with inexpensive mobile robots. J. Comput. Sci. Coll. 16(4), 21–33 (2001)
38.
Zurück zum Zitat Hashimoto, T., Kato, N., Kobayashi, H.: Development of educational system with the android robot SAYA and evaluation. Int. J. Adv. Robot. Syst. 8(3), 28 (2011)CrossRef Hashimoto, T., Kato, N., Kobayashi, H.: Development of educational system with the android robot SAYA and evaluation. Int. J. Adv. Robot. Syst. 8(3), 28 (2011)CrossRef
39.
Zurück zum Zitat Saerbeck, M., Schut, T., Bartneck, C., Janse, M.D.: Expressive robots in education: varying the degree of social supportive behavior of a robotic tutor. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 1613–1622. ACM (2010) Saerbeck, M., Schut, T., Bartneck, C., Janse, M.D.: Expressive robots in education: varying the degree of social supportive behavior of a robotic tutor. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 1613–1622. ACM (2010)
40.
Zurück zum Zitat Srinivasan, V., Takayama, L.: Help me please: robot politeness strategies for soliciting help from humans. In: Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, pp. 4945–4955. ACM (2016) Srinivasan, V., Takayama, L.: Help me please: robot politeness strategies for soliciting help from humans. In: Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, pp. 4945–4955. ACM (2016)
41.
Zurück zum Zitat Hayes, C.C., Miller, C.A.: Human-Computer Etiquette: Cultural Expectations and the Design Implications they Place on Computers and Technology. CRC Press, Boca Raton (2010)CrossRef Hayes, C.C., Miller, C.A.: Human-Computer Etiquette: Cultural Expectations and the Design Implications they Place on Computers and Technology. CRC Press, Boca Raton (2010)CrossRef
42.
Zurück zum Zitat Parasuraman, R., Miller, C.A.: Trust and etiquette in high-criticality automated systems. Commun. ACM 47(4), 51–55 (2004)CrossRef Parasuraman, R., Miller, C.A.: Trust and etiquette in high-criticality automated systems. Commun. ACM 47(4), 51–55 (2004)CrossRef
43.
Zurück zum Zitat de Visser, E.J., Pak, R., Shaw, T.H.: From ‘automation’ to ‘autonomy’: the importance of trust repair in human-machine interaction, Ergonomics 61, 1409–1427 (2018)CrossRef de Visser, E.J., Pak, R., Shaw, T.H.: From ‘automation’ to ‘autonomy’: the importance of trust repair in human-machine interaction, Ergonomics 61, 1409–1427 (2018)CrossRef
44.
Zurück zum Zitat de Visser, E.J., et al.: Almost human: anthropomorphism increases trust resilience in cognitive agents. J. Exp. Psychol.: Appl. 22(3), 331 (2016) de Visser, E.J., et al.: Almost human: anthropomorphism increases trust resilience in cognitive agents. J. Exp. Psychol.: Appl. 22(3), 331 (2016)
45.
Zurück zum Zitat Jackson, R.B., Wen, R., Williams, T.: Tact in noncompliance: the need for pragmatically apt responses to unethical commands (2019) Jackson, R.B., Wen, R., Williams, T.: Tact in noncompliance: the need for pragmatically apt responses to unethical commands (2019)
Metadaten
Titel
Robot Authority in Human-Machine Teams: Effects of Human-Like Appearance on Compliance
verfasst von
Kerstin S. Haring
Ariana Mosley
Sarah Pruznick
Julie Fleming
Kelly Satterfield
Ewart J. de Visser
Chad C. Tossell
Gregory Funke
Copyright-Jahr
2019
DOI
https://doi.org/10.1007/978-3-030-21565-1_5

Neuer Inhalt