Law and Ethics for Robot Soldiers
American University - Washington College of Law; Stanford University - The Hoover Institution on War, Revolution and Peace; Brookings Institution - Governance Studies
Matthew C. Waxman
Columbia Law School
April 5, 2012
Policy Review, 2012
Columbia Public Law Research Paper No. 12-313
American University, WCL Research Paper No. 2012-32
Lethal autonomous machines will inevitably enter the future battlefield — but they will do so incrementally, one small step at a time. The combination of inevitable and incremental development raises not only complex strategic and operational questions but also profound legal and ethical ones. The inevitability of these technologies comes from both supply-side and demand-side factors. Advances in sensor and computational technologies will supply “smarter” machines that can be programmed to kill or destroy, while the increasing tempo of military operations and political pressures to protect one’s own personnel and civilian persons and property will demand continuing research, development, and deployment.
The process will be incremental because non-lethal robotic systems (already proliferating on the battlefield) can be fitted in their successive generations with both self-defensive and offensive technologies. As lethal systems are initially deployed, they may include humans in the decision-making loop, at least as a fail-safe — but as both the decision-making power of machines and the tempo of operations potentially increase, that human role will likely but slowly diminish. Recognizing the inevitable but incremental evolution of these technologies is key to addressing the legal and ethical dilemmas associated with them — U.S. policy for resolving those dilemmas should be built on these assumptions.
The certain yet gradual development and deployment of these systems, as well as the humanitarian advantages created by the precision of some systems, make some proposed responses — such as prohibitory treaties — unworkable as well as ethically questionable. Those features also make it imperative, though, that the United States resist its own impulses toward secrecy and reticence with respect to military technologies, recognizing that the interests those tendencies serve are counterbalanced here by interests in shaping the normative terrain — the contours of international law as well as international expectations about appropriate conduct — on which it and others will operate militarily as technology evolves. Just as development of autonomous weapon systems will be incremental, so too will development of norms about acceptable systems and uses be incremental. The United States must act, however, before international expectations about these technologies harden around the views of those who would impose unrealistic, ineffective or dangerous prohibitions — or those who would prefer few or no constraints at all.
(Note: This paper has been significantly revised for publication in 2013. The finalized version is available as "Law and Ethics for Autonomous Weapon Systems: Why a Ban Won't Work and How the Laws of War Can," The Hoover Institution National Security and Law Essay Series 2013, available at SSRN as http://ssrn.com/abstract=2250126.)
Note: (Annotated version of an essay to appear in a general interest policy journal that does not use footnotes in its articles; sources have been added here for scholarly convenience.)
Number of Pages in PDF File: 20
Keywords: robots, robotics, autonomous lethal weapons, automation, autonomy, unmanned aerial vehicles, UAVs), laws of war, proportionality, distinction, discrimination, weapons, treaties, arms control, OODA
JEL Classification: K10, K33
Date posted: April 26, 2012 ; Last revised: September 5, 2014
© 2015 Social Science Electronic Publishing, Inc. All Rights Reserved.
This page was processed by apollo1 in 1.250 seconds