Browse Items by:
Sign on to:
Please use this identifier to cite or link to this item:
Can human and artificial agents share an autonomy, categorical imperative-based ethics and “moral” selfhood?
autonomous artificial intelligence
ethics for AI
Uniwersytet im. A. Mickiewicza w Poznaniu Wydział Nauk Społecznych, Instytut Filozofii UAM
Filozofia Publiczna i Edukacja Demokratyczna, 2017, Tom 6, Nr 2, s. 169-208.
AI designers endeavour to improve ‘autonomy’ in artificial intelligent devices, as recent developments show. This chapter firstly argues against attributing metaphysical attitudes to AI and, simultaneously, in favor of improving autonomous AI which has been enabled to respect autonomy in human agents. This seems to be the only responsible way of making further advances in the field of autonomous social AI. Let us examine what is meant by claims such as designing our artificial alter egos and sharing moral selves with artificial humanoid devices as well as providing autonomous AI with an ethical framework modelled upon the core aspects of moral selfhood, e.g., making decisions which are based on autonomous law-giving, in Kantian terms.
National Science Centre (NCN, Kraków, Poland), grant OPUS 9, no 2015/17/B/HS1/02381.
Appears in Collections:
Filozofia Publiczna i Edukacja Demokratyczna, 2017, Tom 6, Nr 2
Files in This Item:
07_Ewa_Nowak - Can human and artificial agents share an autonomy categorical imperative-based ethics and moral selfhood.pdf
Show full item record
This item is licensed under a
Creative Commons License