Mammalian Value Systems

Informatica 41(3) (2017)

12 Pages Posted: 31 May 2017 Last revised: 23 Jan 2019

See all articles by Gopal Sarma

Gopal Sarma

School of Medicine, Emory University

Nick Hay

University of California, Berkeley - Computer Science Division

Date Written: May 26, 2017

Abstract

Characterizing human values is a topic deeply interwoven with the sciences, humanities, political philosophy, art, and many other human endeavors. In recent years, a number of thinkers have argued that accelerating trends in computer science, cognitive science, and related disciplines foreshadow the creation of intelligent machines which meet and ultimately surpass the cognitive abilities of human beings, thereby entangling an understanding of human values with future technological development. Contemporary research accomplishments suggest increasingly sophisticated AI systems becoming widespread and responsible for managing many aspects of the modern world, from preemptively planning users’ travel schedules and logistics, to fully autonomous vehicles, to domestic robots assisting in daily living. The extrapolation of these trends has been most forcefully described in the context of a hypothetical “intelligence explosion,” in which the capabilities of an intelligent software agent would rapidly increase due to the presence of feedback loops unavailable to biological organisms. The possibility of superintelligent agents, or simply the widespread deployment of sophisticated, autonomous AI systems, highlights an important theoretical problem: the need to separate the cognitive and rational capacities of an agent from the fundamental goal structure, or value system, which constrains and guides the agent’s actions. The “value alignment problem” is to specify a goal structure for autonomous agents compatible with human values. In this brief article, we suggest that ideas from affective neuroscience and related disciplines aimed at characterizing neurological and behavioral universals in the mammalian class provide important conceptual foundations relevant to describing human values. We argue that the notion of “mammalian value systems” points to a potential avenue for fundamental research in AI safety and AI ethics.

Keywords: AI Safety, Affective Neuroscience, Formal Theory of Values, Orthogonality Thesis, Comparative Neuroanatomy, Evolutionary Psychology, Intelligence Explosion, Value Alignment, Value Learning

Suggested Citation

Sarma, Gopal and Hay, Nick, Mammalian Value Systems (May 26, 2017). Informatica 41(3) (2017), Available at SSRN: https://ssrn.com/abstract=2975399 or http://dx.doi.org/10.2139/ssrn.2975399

Gopal Sarma (Contact Author)

School of Medicine, Emory University ( email )

201 Dowman Drive
Atlanta, GA 30322
United States

Nick Hay

University of California, Berkeley - Computer Science Division ( email )

Berkeley, CA 94720-1712
United States

Do you have a job opening that you would like to promote on SSRN?

Paper statistics

Downloads
90
Abstract Views
828
Rank
625,450
PlumX Metrics