Accountability in Artificial Intelligence: What It Is and How It Works
AI & Society: Journal of Knowledge, Culture and Communication - Springer https://doi.org/10.1007/s00146-023-01635-y
21 Pages Posted: 9 Aug 2022 Last revised: 7 Feb 2023
Date Written: August 3, 2022
Accountability is a cornerstone of the governance of artificial intelligence (AI). However, it is often defined too imprecisely because its multifaceted nature and the sociotechnical structure of AI systems imply a variety of values, practices, and measures to which accountability in AI can refer. We address this lack of clarity by defining accountability in terms of answerability, identifying three conditions of possibility (authority recognition, interrogation, and limitation of power), and an architecture of seven features (context, range, agent, forum, standards, process, and implications). We analyse this architecture through four accountability goals (compliance, report, oversight, and enforcement). We argue that these goals are often complementary and that policy-makers emphasise or prioritise some over others depending on the proactive or reactive use of accountability and the missions of AI governance.
Keywords: AI, accountability, AI Act, governance, policy
Suggested Citation: Suggested Citation