Show simple item record

dc.contributor.authorPerets, Oriel
dc.contributor.authorStagno, Emanuela
dc.contributor.authorBen Yehuda, Eyal
dc.contributor.authorMcNichol, Megan
dc.contributor.authorCeli, Leo
dc.contributor.authorRappoport, Nadav
dc.contributor.authorDorotic, Matilda
dc.date.accessioned2025-09-16T16:57:08Z
dc.date.available2025-09-16T16:57:08Z
dc.date.issued2024-08-05
dc.identifier.issn2157-6904
dc.identifier.urihttps://hdl.handle.net/1721.1/162661
dc.description.abstractBiases inherent in electronic health records (EHRs), a common data source for training medical AI models, may exacerbate health inequities and hinder the adoption of ethical, responsible AI in healthcare. These biases originate from various sources, including implicit clinician biases, data collection and labeling practices, medical devices, and tools used for data processing. Such biases undermine data reliability, influence clinical decisions, and worsen healthcare disparities. When EHR data is used to develop data-driven solutions, biases can further propagate, creating systems that perpetuate inequities. This scoping review categorizes the primary sources of bias in EHRs. We conducted a literature search on PubMed and Web of Science (January 19, 2023) for English-language studies published between 2016 and 2023, following the PRISMA methodology. From 430 initial papers, 27 duplicates were removed, and 403 studies were screened for eligibility. After title, abstract, and full-text reviews, 116 articles were included in the final analysis. Existing studies often focus on isolated biases in EHRs but lack a comprehensive taxonomy. To address this gap, we propose a systematic classification framework encompassing six key sources of bias: (a) biases from prior clinical trials; (b) data-related biases, such as missing or incomplete information; (c) implicit clinician bias; (d) referral and admission bias; (e) diagnosis or risk disparity biases; and (f) biases in medical devices and algorithms. This taxonomy, outlined in Table 1, provides a foundation for evaluating and addressing these issues. While machine learning has transformative potential in healthcare, its effectiveness depends on the integrity of its inputs. Current evidence predominantly addresses data-related biases, with less attention to human or device-related biases, which are often anecdotal or underexplored. For example, racial biases in EHRs are well-documented, but gender-related, sexual orientation, and socially induced biases remain less studied. Compounding biases from these diverse sources can significantly impact AI recommendations, clinical decisions, and patient outcomes. Our review underscores the prevalence of data, human, and machine biases in healthcare and their role in amplifying disparities. To mitigate these challenges, we recommend adopting a ?bias-in-mind? approach when designing data-driven solutions, along with developing safeguards and generating more empirical evidence on bias impacts. This holistic understanding is essential for ensuring equitable and reliable AI applications in healthcare.en_US
dc.publisherACMen_US
dc.relation.isversionofhttp://dx.doi.org/10.1145/3757924en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titleInherent Bias in Electronic Health Records: A Scoping Review of Sources of Biasen_US
dc.typeArticleen_US
dc.identifier.citationOriel Perets, Emanuela Stagno, Eyal Ben Yehuda, Megan McNichol, Leo Anthony Celi, Nadav Rappoport, and Matilda Dorotic. 2025. Inherent Bias in Electronic Health Records: A Scoping Review of Sources of Bias. ACM Trans. Intell. Syst. Technol. Just Accepted (August 2025).en_US
dc.contributor.departmentInstitute for Medical Engineering and Scienceen_US
dc.relation.journalACM Transactions on Intelligent Systems and Technologyen_US
dc.identifier.mitlicensePUBLISHER_POLICY
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2025-09-01T07:57:42Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2025-09-01T07:57:43Z
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record