Show simple item record

dc.contributor.authorDiamant, Nathaniel
dc.contributor.authorReinertsen, Erik
dc.contributor.authorSong, Steven
dc.contributor.authorAguirre, Aaron D
dc.contributor.authorStultz, Collin M
dc.contributor.authorBatra, Puneet
dc.date.accessioned2022-07-20T17:33:16Z
dc.date.available2022-07-20T17:33:16Z
dc.date.issued2022
dc.identifier.urihttps://hdl.handle.net/1721.1/143901
dc.description.abstract<jats:p>Supervised machine learning applications in health care are often limited due to a scarcity of labeled training data. To mitigate the effect of small sample size, we introduce a pre-training approach, <jats:bold>P</jats:bold>atient <jats:bold>C</jats:bold>ontrastive <jats:bold>L</jats:bold>earning of <jats:bold>R</jats:bold>epresentations (PCLR), which creates latent representations of electrocardiograms (ECGs) from a large number of unlabeled examples using contrastive learning. The resulting representations are expressive, performant, and practical across a wide spectrum of clinical tasks. We develop PCLR using a large health care system with over 3.2 million 12-lead ECGs and demonstrate that training linear models on PCLR representations achieves a 51% performance increase, on average, over six training set sizes and four tasks (sex classification, age regression, and the detection of left ventricular hypertrophy and atrial fibrillation), relative to training neural network models from scratch. We also compared PCLR to three other ECG pre-training approaches (supervised pre-training, unsupervised pre-training with an autoencoder, and pre-training using a contrastive multi ECG-segment approach), and show significant performance benefits in three out of four tasks. We found an average performance benefit of 47% over the other models and an average of a 9% performance benefit compared to best model for each task. We release PCLR to enable others to extract ECG representations at <jats:ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:href="https://github.com/broadinstitute/ml4h/tree/master/model_zoo/PCLR" xlink:type="simple">https://github.com/broadinstitute/ml4h/tree/master/model_zoo/PCLR</jats:ext-link>.</jats:p>en_US
dc.language.isoen
dc.publisherPublic Library of Science (PLoS)en_US
dc.relation.isversionof10.1371/JOURNAL.PCBI.1009862en_US
dc.rightsCreative Commons Attribution 4.0 International licenseen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourcePLoSen_US
dc.titlePatient contrastive learning: A performant, expressive, and practical approach to electrocardiogram modelingen_US
dc.typeArticleen_US
dc.identifier.citationDiamant, Nathaniel, Reinertsen, Erik, Song, Steven, Aguirre, Aaron D, Stultz, Collin M et al. 2022. "Patient contrastive learning: A performant, expressive, and practical approach to electrocardiogram modeling." PLoS Computational Biology, 18 (2).
dc.contributor.departmentMassachusetts Institute of Technology. Research Laboratory of Electronics
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
dc.contributor.departmentHarvard University--MIT Division of Health Sciences and Technology
dc.relation.journalPLoS Computational Biologyen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2022-07-20T17:03:37Z
dspace.orderedauthorsDiamant, N; Reinertsen, E; Song, S; Aguirre, AD; Stultz, CM; Batra, Pen_US
dspace.date.submission2022-07-20T17:03:38Z
mit.journal.volume18en_US
mit.journal.issue2en_US
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record