Rethinking Explainable Machines: The GDPR's 'Right to Explanation' Debate and the Rise of Algorithmic Audits in Enterprise

50 Pages Posted: 22 Mar 2018

See all articles by Bryan Casey

Bryan Casey

Stanford Law School

Ashkon Farhangi

Stanford University - Stanford Codex Center

Roland Vogl

Stanford Law School

Date Written: February 19, 2018

Abstract

The public debate surrounding the General Data Protection Regulation’s (“GDPR”) “right to explanation” has sparked a global conversation of profound social and economic significance. But from a practical perspective, the debate’s participants have gotten way ahead of themselves. In their search for a revolutionary new data protection within the provisions of a single chapter of the GDPR, many prominent contributors to the debate have lost sight of the most revolutionary change ushered in by the Regulation: the sweeping new enforcement powers given to European data protection authorities (“DPAs”) by Chapters 6 and 8 of the Regulation. Unlike the 1995 Data Protection Directive that it will replace, the GDPR’s potent new investigatory, advisory, corrective, and punitive powers granted by Chapters 6 and 8 will render DPAs de facto interpretive authorities of the Regulation’s controversial “right to explanation.” Now that the DPAs responsible for enforcing the right have officially weighed in, this Article argues that at least one matter of fierce public debate can be laid to rest. The GDPR provides an unambiguous “right to explanation” with sweeping legal implications for the design, prototyping, field testing, and deployment of automated data processing systems. While the protections enshrined within the right may not mandate transparency in the form of a complete individualized explanation, a holistic understanding of the Regulation’s interpretation by DPAs reveals that the right’s true power derives from its synergistic effects when combined with the algorithmic auditing and “data protection by design” methodologies codified by the Regulation’s subsequent chapters. Accordingly, this Article predicts that algorithmic auditing and “data protection by design” practices will likely become the new gold standard for enterprises deploying machine learning systems both inside and outside of the EU bloc.

Keywords: Automated, Decision, Decision-Making, Profiling, Data, Processing, Protection, Explanation, XAI, Explicability, Explainable, Interpretability, Machine, Learning, AI, Artificial, Intelligence, Algorithm

Suggested Citation

Casey, Bryan and Farhangi, Ashkon and Vogl, Roland, Rethinking Explainable Machines: The GDPR's 'Right to Explanation' Debate and the Rise of Algorithmic Audits in Enterprise (February 19, 2018). Berkeley Technology Law Journal, Forthcoming. Available at SSRN: https://ssrn.com/abstract=3143325

Bryan Casey (Contact Author)

Stanford Law School ( email )

559 Nathan Abbott Way
Stanford, CA 94305-8610
United States

Ashkon Farhangi

Stanford University - Stanford Codex Center ( email )

559 Nathan Abbott Way
Stanford, CA 94305-8610
United States

Roland Vogl

Stanford Law School ( email )

559 Nathan Abbott Way
Stanford, CA 94305-8610
United States

Register to save articles to
your library

Register

Paper statistics

Downloads
466
rank
58,143
Abstract Views
1,669
PlumX Metrics