Show simple item record

dc.contributor.authorCasper, Stephen
dc.contributor.authorEzell, Carson
dc.contributor.authorSiegmann, Charlotte
dc.contributor.authorKolt, Noam
dc.contributor.authorCurtis, Taylor Lynn
dc.contributor.authorBucknall, Benjamin
dc.contributor.authorHaupt, Andreas
dc.contributor.authorWei, Kevin
dc.contributor.authorScheurer, Jérémy
dc.contributor.authorHobbhahn, Marius
dc.contributor.authorSharkey, Lee
dc.contributor.authorKrishna, Satyapriya
dc.contributor.authorVon Hagen, Marvin
dc.contributor.authorAlberti, Silas
dc.contributor.authorChan, Alan
dc.contributor.authorSun, Qinyi
dc.contributor.authorGerovitch, Michael
dc.contributor.authorBau, David
dc.contributor.authorTegmark, Max
dc.contributor.authorKrueger, David
dc.contributor.authorHadfield-Menell, Dylan
dc.date.accessioned2024-07-24T17:19:49Z
dc.date.available2024-07-24T17:19:49Z
dc.date.issued2024-06-03
dc.identifier.isbn979-8-4007-0450-5
dc.identifier.urihttps://hdl.handle.net/1721.1/155783
dc.descriptionFAccT ’24, June 03–06, 2024, Rio de Janeiro, Brazilen_US
dc.description.abstractExternal audits of AI systems are increasingly recognized as a key mechanism for AI governance. The effectiveness of an audit, however, depends on the degree of access granted to auditors. Recent audits of state-of-the-art AI systems have primarily relied on black-box access, in which auditors can only query the system and observe its outputs. However, white-box access to the system’s inner workings (e.g., weights, activations, gradients) allows an auditor to perform stronger attacks, more thoroughly interpret models, and conduct fine-tuning. Meanwhile, outside-the-box access to training and deployment information (e.g., methodology, code, documentation, data, deployment details, findings from internal evaluations) allows auditors to scrutinize the development process and design more targeted evaluations. In this paper, we examine the limitations of black-box audits and the advantages of white- and outside-the-box audits. We also discuss technical, physical, and legal safeguards for performing these audits with minimal security risks. Given that different forms of access can lead to very different levels of evaluation, we conclude that (1) transparency regarding the access and methods used by auditors is necessary to properly interpret audit results, and (2) white- and outside-the-box access allow for substantially more scrutiny than black-box access alone.en_US
dc.publisherACM|The 2024 ACM Conference on Fairness, Accountability, and Transparencyen_US
dc.relation.isversionof10.1145/3630106.3659037en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titleBlack-Box Access is Insufficient for Rigorous AI Auditsen_US
dc.typeArticleen_US
dc.identifier.citationCasper, Stephen, Ezell, Carson, Siegmann, Charlotte, Kolt, Noam, Curtis, Taylor Lynn et al. 2024. "Black-Box Access is Insufficient for Rigorous AI Audits."
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
dc.contributor.departmentMassachusetts Institute of Technology. Department of Economics
dc.contributor.departmentMassachusetts Institute of Technology. Center for Collective Intelligence
dc.contributor.departmentMassachusetts Institute of Technology. Department of Physics
dc.identifier.mitlicensePUBLISHER_CC
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2024-07-01T07:56:42Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2024-07-01T07:56:43Z
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record