Accelerating AI

26 Pages Posted: 25 Apr 2010

Date Written: April 21, 2010

Abstract

This essay argues for government support for “friendly AI” - a kind of artificial intelligence that will not endanger humans. AI is not, contrary to some theorists, necessarily a threat to humans, because AI will not necessarily be anthropomorphic and possess an all-too-human will to power. Government support for friendly AI makes it more like that friendly AI will emerge before other more dangerous kinds. Moreover, such support is also justified because the acceleration of AI will aid in the analysis of the risks and benefits created by the many other kinds of technology that are also accelerating - from nanotechnology to biotechnology. Only AI of all the many accelerating technologies of our present day helps perform this crucial function. Alternative approaches to AI - relinquishment or regulation of kinds of AI that might be deemed harmful - are infeasible. Given that AI is so central to modern military power, relinquishing or inhibiting AI would empower the worst nations on earth.

Keywords: Artificial Intelligence, Technology, Weapons, Information

JEL Classification: K19, K39

Suggested Citation

McGinnis, John, Accelerating AI (April 21, 2010). Northwestern University Law Review, Forthcoming, Northwestern Public Law Research Paper No. 10-12, Available at SSRN: https://ssrn.com/abstract=1593851 or http://dx.doi.org/10.2139/ssrn.1593851

John McGinnis (Contact Author)

Northwestern University - Pritzker School of Law ( email )

375 E. Chicago Ave
Chicago, IL 60611
United States
312-503-3235 (Phone)

Do you have a job opening that you would like to promote on SSRN?

Paper statistics

Downloads
659
Abstract Views
5,097
Rank
80,950
PlumX Metrics