Abstract

http://ssrn.com/abstract=983503
 
 

References (9)



 


 



Artificial Agency, Consciousness, and the Criteria for Moral Agency: What Properties Must an Artificial Agent Have to Be a Moral Agent?


Kenneth Einar Himma


University of Washington - School of Law

April 27, 2007

Ethics and Information Technology, Vol. 11, No. 1, 2009

Abstract:     
A spate of papers have recently appeared raising the issue of the possibility of not only artificial agency, but also artificial moral agency that raises all sorts of substantive questions of moral responsibility. Suppose, for example, that we are epistemically justified in believing that an ICT engineer has designed and produced an ICT that is capable of acting and satisfies the criteria for moral agency. If this ICT turns out to do something bad that no one anticipated, who is morally responsible: the designer, the ICT or some combination of both?

What I wish to attempt to do is to work out some conceptual issues regarding the concepts of agency, natural agency, artificial agency, and moral agency, as well as articulate the criteria for moral agency as a first step towards beginning a consideration of these important questions of professional responsibility. Much of what I take myself to be doing enjoys a consensus in the literature - so much so that many crucial claims on which I rely are taken for granted in such widely-used and respected professional resources as the Stanford Encyclopedia of Philosophy, Routledge Encyclopedia of Philosophy, and the Internet Encyclopedia of Philosophy. I attempt to flesh out some of the implications of some of these well-settled theories with respect to the prerequisites that an ICT must satisfy in order to count as a moral agent accountable for its behavior.

I will begin with analyses of the more basic concepts, like that of agency, and work up to an analyses of the more complex concepts, like that of moral agency and rational free agency, subsequently considering the criteria something must satisfy to be accountable for its behavior; all of this will largely be uncontroversial in the literature. I will then argue that the each of the various elements of the necessary conditions for moral agency presupposes consciousness, i.e., the capacity for inner subjective experience like that of pain or, as Nagel puts it, the possession of an internal something-of-which-it is-is-to-be-like. I ultimately conclude that the issue of whether artificial moral agency is possible depends on the issue of whether it is possible for ICTs to be conscious.

Number of Pages in PDF File: 11

Keywords: action, agency, artificial agency, moral agency, moral accountability, consciousness

Accepted Paper Series


Download This Paper

Date posted: May 1, 2007 ; Last revised: March 27, 2014

Suggested Citation

Himma, Kenneth Einar, Artificial Agency, Consciousness, and the Criteria for Moral Agency: What Properties Must an Artificial Agent Have to Be a Moral Agent? (April 27, 2007). Ethics and Information Technology, Vol. 11, No. 1, 2009. Available at SSRN: http://ssrn.com/abstract=983503 or http://dx.doi.org/10.2139/ssrn.983503

Contact Information

Kenneth Einar Himma (Contact Author)
University of Washington - School of Law ( email )
William Gates Hall
Box 353020
Seattle, WA 98195-3020
United States
206-543-4550 (Phone)
Feedback to SSRN


Paper statistics
Abstract Views: 1,399
Downloads: 239
Download Rank: 73,703
References:  9

© 2014 Social Science Electronic Publishing, Inc. All Rights Reserved.  FAQ   Terms of Use   Privacy Policy   Copyright   Contact Us
This page was processed by apollo6 in 0.250 seconds