Skip to main content
Erschienen in:
Buchtitelbild

2018 | OriginalPaper | Buchkapitel

Hybrid Strategies Towards Safe “Self-Aware” Superintelligent Systems

verfasst von : Nadisha-Marie Aliman, Leon Kester

Erschienen in: Artificial General Intelligence

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Against the backdrop of increasing progresses in AI research paired with a rise of AI applications in decision-making processes, security-critical domains as well as in ethically relevant frames, a large-scale debate on possible safety measures encompassing corresponding long-term and short-term issues has emerged across different disciplines. One pertinent topic in this context which has been addressed by various AI Safety researchers is e.g. the AI alignment problem for which no final consensus has been achieved yet. In this paper, we present a multidisciplinary toolkit of AI Safety strategies combining considerations from AI and Systems Engineering as well as from Cognitive Science with a security mindset as often relevant in Cybersecurity. We elaborate on how AGI “Self-awareness” could complement different AI Safety measures in a framework extended by a jointly performed Human Enhancement procedure. Our analysis suggests that this hybrid framework could contribute to undertake the AI alignment problem from a new holistic perspective through security-building synergetic effects emerging thereof and could help to increase the odds of a possible safe future transition towards superintelligent systems.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Adams, S.S., et al.: Mapping the landscape of human-level artificial general intelligence. AI Magaz. 33, 25–41 (2012)CrossRef Adams, S.S., et al.: Mapping the landscape of human-level artificial general intelligence. AI Magaz. 33, 25–41 (2012)CrossRef
3.
Zurück zum Zitat Ashby, M.: Ethical regulators and super-ethical systems. In: Proceedings of the 61st Annual Meeting of the ISSS-2017 Vienna, Austria, vol. 2017 (2017) Ashby, M.: Ethical regulators and super-ethical systems. In: Proceedings of the 61st Annual Meeting of the ISSS-2017 Vienna, Austria, vol. 2017 (2017)
4.
Zurück zum Zitat Ashby, W.R.: An Introduction to Cybernetics. Chapman & Hall Ltd., New York (1961) Ashby, W.R.: An Introduction to Cybernetics. Chapman & Hall Ltd., New York (1961)
5.
Zurück zum Zitat Baars, B.J., Franklin, S.: Consciousness is computational: the LIDA model of global workspace theory. Int. J. Mach. Conscious. 1(01), 23–32 (2009)CrossRef Baars, B.J., Franklin, S.: Consciousness is computational: the LIDA model of global workspace theory. Int. J. Mach. Conscious. 1(01), 23–32 (2009)CrossRef
6.
Zurück zum Zitat Bach, J.: Principles of Synthetic Intelligence PSI: An Architecture of Motivated Cognition, vol. 4. Oxford University Press, Oxford (2009) Bach, J.: Principles of Synthetic Intelligence PSI: An Architecture of Motivated Cognition, vol. 4. Oxford University Press, Oxford (2009)
7.
Zurück zum Zitat Bostrom, N.: Superintelligence: Paths, Dangers, Strategies (2014) Bostrom, N.: Superintelligence: Paths, Dangers, Strategies (2014)
8.
Zurück zum Zitat Brundage, M., et al.: The malicious use of artificial intelligence: forecasting, prevention, and mitigation. arXiv preprint arXiv:1802.07228 (2018) Brundage, M., et al.: The malicious use of artificial intelligence: forecasting, prevention, and mitigation. arXiv preprint arXiv:​1802.​07228 (2018)
9.
Zurück zum Zitat Drumwright, M., Prentice, R., Biasucci, C.: Behavioral ethics and teaching ethical decision making. Decis. Sci. J. Innovative Educ. 13(3), 431–458 (2015)CrossRef Drumwright, M., Prentice, R., Biasucci, C.: Behavioral ethics and teaching ethical decision making. Decis. Sci. J. Innovative Educ. 13(3), 431–458 (2015)CrossRef
10.
Zurück zum Zitat van Foeken, E., Kester, L., Iersel, M.: Real-time common awareness in communication constrained sensor systems. In: Proceedings of 12th International Conference on Information Fusion, FUSION 2009, Seattle, Washington, USA, pp. 118–125, 6–9 July 2009 van Foeken, E., Kester, L., Iersel, M.: Real-time common awareness in communication constrained sensor systems. In: Proceedings of 12th International Conference on Information Fusion, FUSION 2009, Seattle, Washington, USA, pp. 118–125, 6–9 July 2009
11.
Zurück zum Zitat Goertzel, B.: Should humanity build a global AI nanny to delay the singularity until its better understood? J. Conscious. Stud. 19(1–2), 96–111 (2012) Goertzel, B.: Should humanity build a global AI nanny to delay the singularity until its better understood? J. Conscious. Stud. 19(1–2), 96–111 (2012)
12.
Zurück zum Zitat Goertzel, B.: Characterizing human-like consciousness: an integrative approach. Procedia Comput. Sci. 41, 152–157 (2014)CrossRef Goertzel, B.: Characterizing human-like consciousness: an integrative approach. Procedia Comput. Sci. 41, 152–157 (2014)CrossRef
14.
Zurück zum Zitat Kester, L., Ditzel, M.: Maximising effectiveness of distributed mobile observation systems in dynamic situations. In: 2014 17th International Conference on Information Fusion (FUSION), pp. 1–8. IEEE (2014) Kester, L., Ditzel, M.: Maximising effectiveness of distributed mobile observation systems in dynamic situations. In: 2014 17th International Conference on Information Fusion (FUSION), pp. 1–8. IEEE (2014)
15.
Zurück zum Zitat Kester, L.J.H.M., van Willigen, W.H., Jongh, J.D.: Critical headway estimation under uncertainty and non-ideal communication conditions. In: Proceedings of 17th International IEEE Conference on Intelligent Transportation Systems (ITSC), pp. 320–327 (2014) Kester, L.J.H.M., van Willigen, W.H., Jongh, J.D.: Critical headway estimation under uncertainty and non-ideal communication conditions. In: Proceedings of 17th International IEEE Conference on Intelligent Transportation Systems (ITSC), pp. 320–327 (2014)
16.
Zurück zum Zitat Mirkovic, J., et al.: Testing a collaborative DDoS defense in a red team/blue team exercise. IEEE Trans. Comput. 57(8), 1098–1112 (2008)MathSciNetCrossRef Mirkovic, J., et al.: Testing a collaborative DDoS defense in a red team/blue team exercise. IEEE Trans. Comput. 57(8), 1098–1112 (2008)MathSciNetCrossRef
17.
Zurück zum Zitat Morewedge, C.K., Yoon, H., Scopelliti, I., Symborski, C.W., Korris, J.H., Kassam, K.S.: Debiasing decisions: Improved decision making with a single training intervention. Policy Insights Behav. Brain Sci. 2(1), 129–140 (2015)CrossRef Morewedge, C.K., Yoon, H., Scopelliti, I., Symborski, C.W., Korris, J.H., Kassam, K.S.: Debiasing decisions: Improved decision making with a single training intervention. Policy Insights Behav. Brain Sci. 2(1), 129–140 (2015)CrossRef
19.
Zurück zum Zitat Papernot, N., McDaniel, P., Sinha, A., Wellman, M.: Towards the science of security and privacy in machine learning. arXiv preprint arXiv:1611.03814 (2016) Papernot, N., McDaniel, P., Sinha, A., Wellman, M.: Towards the science of security and privacy in machine learning. arXiv preprint arXiv:​1611.​03814 (2016)
20.
Zurück zum Zitat Pistono, F., Yampolskiy, R.V.: Unethical research: how to create a malevolent artificial intelligence. In: Proceedings of 25th International Joint Conference on Artificial Intelligence (IJCAI-16). Ethics for Artificial Intelligence Workshop (AI-Ethics-2016) (2016) Pistono, F., Yampolskiy, R.V.: Unethical research: how to create a malevolent artificial intelligence. In: Proceedings of 25th International Joint Conference on Artificial Intelligence (IJCAI-16). Ethics for Artificial Intelligence Workshop (AI-Ethics-2016) (2016)
21.
Zurück zum Zitat Potapov, A.: Technological singularity: what do we really know? Information 9(4), 99 (2018) Potapov, A.: Technological singularity: what do we really know? Information 9(4), 99 (2018)
22.
Zurück zum Zitat Rajendran, J., Jyothi, V., Karri, R.: Blue team red team approach to hardware trust assessment. In: 2011 IEEE 29th International Conference on Computer Design (ICCD), pp. 285–288. IEEE (2011) Rajendran, J., Jyothi, V., Karri, R.: Blue team red team approach to hardware trust assessment. In: 2011 IEEE 29th International Conference on Computer Design (ICCD), pp. 285–288. IEEE (2011)
23.
Zurück zum Zitat Rege, A.: Incorporating the human element in anticipatory and dynamic cyber defense. In: IEEE International Conference on Cybercrime and Computer Forensic (ICCCF), pp. 1–7. IEEE (2016) Rege, A.: Incorporating the human element in anticipatory and dynamic cyber defense. In: IEEE International Conference on Cybercrime and Computer Forensic (ICCCF), pp. 1–7. IEEE (2016)
24.
Zurück zum Zitat Rege, A., Obradovic, Z., Asadi, N., Singer, B., Masceri, N.: A temporal assessment of cyber intrusion chains using multidisciplinary frameworks and methodologies. In: 2017 International Conference on Cyber Situational Awareness, Data Analytics and Assessment (Cyber SA), pp. 1–7. IEEE (2017) Rege, A., Obradovic, Z., Asadi, N., Singer, B., Masceri, N.: A temporal assessment of cyber intrusion chains using multidisciplinary frameworks and methodologies. In: 2017 International Conference on Cyber Situational Awareness, Data Analytics and Assessment (Cyber SA), pp. 1–7. IEEE (2017)
26.
Zurück zum Zitat Sezer, O., Gino, F., Bazerman, M.H.: Ethical blind spots: explaining unintentional unethical behavior. Curr. Opin. Psychol. 6, 77–81 (2015)CrossRef Sezer, O., Gino, F., Bazerman, M.H.: Ethical blind spots: explaining unintentional unethical behavior. Curr. Opin. Psychol. 6, 77–81 (2015)CrossRef
28.
Zurück zum Zitat Soares, N., Fallenstein, B.: Agent foundations for aligning machine intelligence with human interests: a technical research agenda. In: Callaghan, V., Miller, J., Yampolskiy, R., Armstrong, S. (eds.) The Technological Singularity. TFC, pp. 103–125. Springer, Heidelberg (2017). https://doi.org/10.1007/978-3-662-54033-6_5CrossRef Soares, N., Fallenstein, B.: Agent foundations for aligning machine intelligence with human interests: a technical research agenda. In: Callaghan, V., Miller, J., Yampolskiy, R., Armstrong, S. (eds.) The Technological Singularity. TFC, pp. 103–125. Springer, Heidelberg (2017). https://​doi.​org/​10.​1007/​978-3-662-54033-6_​5CrossRef
29.
Zurück zum Zitat Taylor, J., Yudkowsky, E., LaVictoire, P., Critch, A.: Alignment for advanced machine learning systems. In: Machine Intelligence Research Institute (2016) Taylor, J., Yudkowsky, E., LaVictoire, P., Critch, A.: Alignment for advanced machine learning systems. In: Machine Intelligence Research Institute (2016)
30.
Zurück zum Zitat Tegmark, M.: Life 3.0: Being Human in the Age of Artificial Intelligence. Knopf, New York (2017) Tegmark, M.: Life 3.0: Being Human in the Age of Artificial Intelligence. Knopf, New York (2017)
32.
Zurück zum Zitat Wang, P., Li, X., Hammer, P.: Self in NARS, an AGI system. Front. Robot. AI 5, 20 (2018)CrossRef Wang, P., Li, X., Hammer, P.: Self in NARS, an AGI system. Front. Robot. AI 5, 20 (2018)CrossRef
34.
Zurück zum Zitat Yudkowsky, E.: Cognitive biases potentially affecting judgment of global risks. Glob. Catastrophic Risks 1(86), 13 (2008) Yudkowsky, E.: Cognitive biases potentially affecting judgment of global risks. Glob. Catastrophic Risks 1(86), 13 (2008)
Metadaten
Titel
Hybrid Strategies Towards Safe “Self-Aware” Superintelligent Systems
verfasst von
Nadisha-Marie Aliman
Leon Kester
Copyright-Jahr
2018
DOI
https://doi.org/10.1007/978-3-319-97676-1_1