Accountability in Artificial Intelligence: What It Is and How It Works
AI & SOCIETY, volume 39, issue 4, 2024[10.1007/s00146-023-01635-y]
21 Pages Posted: 9 Aug 2022 Last revised: 30 Nov 2024
Date Written: August 3, 2022
Abstract
Accountability is a cornerstone of the governance of artificial intelligence (AI). However, it is often defined too imprecisely because its multifaceted nature and the sociotechnical structure of AI systems imply a variety of values, practices, and measures to which accountability in AI can refer. We address this lack of clarity by defining accountability in terms of answerability, identifying three conditions of possibility (authority recognition, interrogation, and limitation of power), and an architecture of seven features (context, range, agent, forum, standards, process, and implications). We analyse this architecture through four accountability goals (compliance, report, oversight, and enforcement). We argue that these goals are often complementary and that policy-makers emphasise or prioritise some over others depending on the proactive or reactive use of accountability and the missions of AI governance.
Keywords: AI, accountability, AI Act, governance, policy
Suggested Citation: Suggested Citation