Research Agenda for Sociotechnical Approaches to AI Safety

26 Pages Posted: 7 Mar 2025

See all articles by Samuel Curtis

Samuel Curtis

The Future Society

Ravi Iyer

University of Southern California

Cameron Domenico Kirk-Giannini

Rutgers, The State University of New Jersey

Victoria Krakovna

Future of Life Institute

David Krueger

University of Cambridge

Nathan Lambert

Allen Institute for AI

Bruno Marnette

AI Objectives Institute

Colleen McKenzie

AI Objectives Institute

Julian Michael

New York University (NYU) - Center for Data Science

Evan Miyazono

Atlas Computing

Noyuri Mima

Future University Hakodate

Aviv Ovadya

Thoughtful Technology Project

Luke Thorburn

King's College London; AI & Democracy Foundation

Vehbi Deger Turan

AI Objectives Institute

Date Written: January 14, 2025

Abstract

As the capabilities of AI systems continue to advance, it is increasingly important that we guide the development of these powerful technologies, ensuring they are used for the benefit of society. Existing work analyzing and assessing risks from AI spans a broad and diverse range of perspectives, including some which diverge enough in their motivations and approaches that they disagree on priorities and desired solutions. Yet we find significant overlap among these perspectives' desire for beneficial outcomes from AI deployment, and significant potential for progress towards such outcomes in the examination of that overlap. In this paper we explore one such area of overlap: we discuss areas of AI safety work that could benefit from sociotechnical framings of AI, which view AI systems as embedded in larger sociotechnical systems, and which explore the potential risks and benefits of AI not just as aspects of these new tools, but as possibilities for the complex interactions between humans and our technologies. We present a collection of proposals we believe to be promising directions for including sociotechnical approaches in the pursuit of safe and beneficial AI, demonstrating the potential value of such approaches in addressing the harms, risks, and benefits of current and future AI systems.

Keywords: AI, Artificial Intelligence, AI Ethics, AI Safety, Science & Technology Studies, AI Governance, sociotechnical, research, AI systems, social welfare, interpretability, RLHF

Suggested Citation

Curtis, Samuel and Iyer, Ravi and Domenico Kirk-Giannini, Cameron and Krakovna, Victoria and Krueger, David and Lambert, Nathan and Marnette, Bruno and McKenzie, Colleen and Michael, Julian and Miyazono, Evan and Mima, Noyuri and Ovadya, Aviv and Thorburn, Luke and Turan, Vehbi Deger, Research Agenda for Sociotechnical Approaches to AI Safety (January 14, 2025). Available at SSRN: https://ssrn.com/abstract=5097286 or http://dx.doi.org/10.2139/ssrn.5097286

Samuel Curtis

The Future Society ( email )

Ravi Iyer

University of Southern California ( email )

Cameron Domenico Kirk-Giannini

Rutgers, The State University of New Jersey ( email )

Victoria Krakovna

Future of Life Institute ( email )

David Krueger

University of Cambridge ( email )

Nathan Lambert

Allen Institute for AI

Bruno Marnette

AI Objectives Institute ( email )

Colleen McKenzie (Contact Author)

AI Objectives Institute ( email )

Julian Michael

New York University (NYU) - Center for Data Science

60 5th Ave
7th Floor
New York, NY 10011
United States

Evan Miyazono

Atlas Computing ( email )

Noyuri Mima

Future University Hakodate ( email )

Aviv Ovadya

Thoughtful Technology Project

Luke Thorburn

King's College London ( email )

United Kingdom

AI & Democracy Foundation

United States

Vehbi Deger Turan

AI Objectives Institute ( email )

Do you have a job opening that you would like to promote on SSRN?

Paper statistics

Downloads
40
Abstract Views
121
PlumX Metrics