Show simple item record

dc.contributor.authorGameiro, Rodrigo R.
dc.contributor.authorWoite, Naira L.
dc.contributor.authorSauer, Christopher M.
dc.contributor.authorHao, Sicheng
dc.contributor.authorFernandes, Chrystinne O.
dc.contributor.authorPremo, Anna E.
dc.contributor.authorTeixeira, Alice R.
dc.contributor.authorResli, Isabelle
dc.contributor.authorWong, An-Kwok I.
dc.contributor.authorCeli, Leo A.
dc.date.accessioned2025-03-03T17:50:57Z
dc.date.available2025-03-03T17:50:57Z
dc.date.issued2025-02-04
dc.identifier.urihttps://hdl.handle.net/1721.1/158286
dc.description.abstractBackground The deployment of Artificial Intelligence (AI) in healthcare has the potential to transform patient care through improved diagnostics, personalized treatment plans, and more efficient resource management. However, the effectiveness and fairness of AI are critically dependent on the data it learns from. Biased datasets can lead to AI outputs that perpetuate disparities, particularly affecting social minorities and marginalized groups. Objective This paper introduces the “Data Artifacts Glossary”, a dynamic, open-source framework designed to systematically document and update potential biases in healthcare datasets. The aim is to provide a comprehensive tool that enhances the transparency and accuracy of AI applications in healthcare and contributes to understanding and addressing health inequities. Methods Utilizing a methodology inspired by the Delphi method, a diverse team of experts conducted iterative rounds of discussions and literature reviews. The team synthesized insights to develop a comprehensive list of bias categories and designed the glossary’s structure. The Data Artifacts Glossary was piloted using the MIMIC-IV dataset to validate its utility and structure. Results The Data Artifacts Glossary adopts a collaborative approach modeled on successful open-source projects like Linux and Python. Hosted on GitHub, it utilizes robust version control and collaborative features, allowing stakeholders from diverse backgrounds to contribute. Through a rigorous peer review process managed by community members, the glossary ensures the continual refinement and accuracy of its contents. The implementation of the Data Artifacts Glossary with the MIMIC-IV dataset illustrates its utility. It categorizes biases, and facilitates their identification and understanding. Conclusion The Data Artifacts Glossary serves as a vital resource for enhancing the integrity of AI applications in healthcare by providing a mechanism to recognize and mitigate dataset biases before they impact AI outputs. It not only aids in avoiding bias in model development but also contributes to understanding and addressing the root causes of health disparities.en_US
dc.publisherBioMed Centralen_US
dc.relation.isversionofhttps://doi.org/10.1186/s12929-024-01106-6en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceBioMed Centralen_US
dc.titleThe Data Artifacts Glossary: a community-based repository for bias on health datasetsen_US
dc.typeArticleen_US
dc.identifier.citationGameiro, R.R., Woite, N.L., Sauer, C.M. et al. The Data Artifacts Glossary: a community-based repository for bias on health datasets. J Biomed Sci 32, 14 (2025).en_US
dc.contributor.departmentHarvard--MIT Program in Health Sciences and Technology. Laboratory for Computational Physiologyen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Urban Studies and Planningen_US
dc.relation.journalJournal of Biomedical Scienceen_US
dc.identifier.mitlicensePUBLISHER_CC
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2025-02-13T10:17:41Z
dc.language.rfc3066en
dc.rights.holderThe Author(s)
dspace.date.submission2025-02-13T10:17:41Z
mit.journal.volume32en_US
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record