dc.contributor.author | Cen, Sarah | |
dc.contributor.author | Alur, Rohan | |
dc.date.accessioned | 2024-11-21T17:38:48Z | |
dc.date.available | 2024-11-21T17:38:48Z | |
dc.date.issued | 2024-10-29 | |
dc.identifier.isbn | 979-8-4007-1222-7 | |
dc.identifier.uri | https://hdl.handle.net/1721.1/157655 | |
dc.description | EAAMO ’24, October 29–31, 2024, San Luis Potosi, Mexico | en_US |
dc.description.abstract | Artificial intelligence (AI) is increasingly intervening in our lives, raising widespread concern about its unintended and undeclared side effects. These developments have brought attention to the problem of AI auditing: the systematic evaluation and analysis of an AI system, its development, and its behavior relative to a set of predetermined criteria. Auditing can take many forms, including pre-deployment risk assessments, ongoing monitoring, and compliance testing. It plays a critical role in providing assurances to various AI stakeholders, from developers to end users. Audits may, for instance, be used to verify that an algorithm complies with the law, is consistent with industry standards, and meets the developer’s claimed specifications. However, AI developers and companies will rarely grant auditors unfettered access to their systems.
In this work, we examine a key consideration in AI auditing: what type of access to an AI system is needed to perform a meaningful audit? Addressing this question has direct policy relevance, as it can inform AI audit guidelines and requirements. We begin by discussing the factors that auditors balance when determining the appropriate type of access, and unpack the benefits and drawbacks of four types of access. We conclude that, at minimum, black-box access—providing query access to a model without exposing its internal implementation—should be granted to auditors. In particular, we argue that black-box access effectively balances concerns related to proprietary technology, data privacy, audit standardization, and audit efficiency. We then suggest a framework for determining how much further access (on top of black-box access) to provide to auditors. We show that auditing can be cast as a natural hypothesis test and argue that this framing provides clear and interpretable guidance on the implementation of AI audits. In particular, we draw parallels between aspects of hypothesis testing and those of legal procedure, such as legal presumption and burden of proof. As a result, hypothesis testing provides an approach to AI auditing that is both interpretable and effective, offering a potential path forward despite the challenges posed by AI’s opacity. | en_US |
dc.publisher | ACM|Equity and Access in Algorithms, Mechanisms, and Optimization | en_US |
dc.relation.isversionof | https://doi.org/10.1145/3689904.3694711 | en_US |
dc.rights | Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use. | en_US |
dc.source | Association for Computing Machinery | en_US |
dc.title | From Transparency to Accountability and Back | en_US |
dc.type | Article | en_US |
dc.identifier.citation | Cen, Sarah and Alur, Rohan. 2024. "From Transparency to Accountability and Back." | |
dc.contributor.department | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science | en_US |
dc.identifier.mitlicense | PUBLISHER_POLICY | |
dc.eprint.version | Final published version | en_US |
dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
dc.date.updated | 2024-11-01T07:54:41Z | |
dc.language.rfc3066 | en | |
dc.rights.holder | The author(s) | |
dspace.date.submission | 2024-11-01T07:54:42Z | |
mit.license | PUBLISHER_POLICY | |
mit.metadata.status | Authority Work and Publication Information Needed | en_US |